var/home/core/zuul-output/0000755000175000017500000000000015112102366014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112114315015463 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005135645115112114307017702 0ustar rootrootNov 27 17:09:23 crc systemd[1]: Starting Kubernetes Kubelet... Nov 27 17:09:23 crc restorecon[4688]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:23 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 17:09:24 crc restorecon[4688]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 27 17:09:25 crc kubenswrapper[4809]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 17:09:25 crc kubenswrapper[4809]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 27 17:09:25 crc kubenswrapper[4809]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 17:09:25 crc kubenswrapper[4809]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 17:09:25 crc kubenswrapper[4809]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 27 17:09:25 crc kubenswrapper[4809]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.188600 4809 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.196947 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.196984 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.196992 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.196999 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197010 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197021 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197026 4809 feature_gate.go:330] unrecognized feature gate: Example Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197032 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197039 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197046 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197055 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197074 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197080 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197085 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197091 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197096 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197101 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197106 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197111 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197116 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197121 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197127 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197133 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197138 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197143 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197147 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197152 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197157 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197162 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197167 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197172 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197177 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197183 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197188 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197193 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197199 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197205 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197210 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197215 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197220 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197225 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197231 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197236 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197244 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197248 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197253 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197258 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197264 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197268 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197273 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197278 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197282 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197287 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197292 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197297 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197301 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197306 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197310 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197315 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197320 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197324 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197329 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197334 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197338 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197344 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197348 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197353 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197357 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197364 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197370 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.197375 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199774 4809 flags.go:64] FLAG: --address="0.0.0.0" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199802 4809 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199814 4809 flags.go:64] FLAG: --anonymous-auth="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199824 4809 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199836 4809 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199844 4809 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199855 4809 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199866 4809 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199874 4809 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199882 4809 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199890 4809 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199898 4809 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199908 4809 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199916 4809 flags.go:64] FLAG: --cgroup-root="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199923 4809 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199930 4809 flags.go:64] FLAG: --client-ca-file="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199937 4809 flags.go:64] FLAG: --cloud-config="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199944 4809 flags.go:64] FLAG: --cloud-provider="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199951 4809 flags.go:64] FLAG: --cluster-dns="[]" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199960 4809 flags.go:64] FLAG: --cluster-domain="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199967 4809 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199974 4809 flags.go:64] FLAG: --config-dir="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199980 4809 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199989 4809 flags.go:64] FLAG: --container-log-max-files="5" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.199999 4809 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200006 4809 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200013 4809 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200021 4809 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200029 4809 flags.go:64] FLAG: --contention-profiling="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200036 4809 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200043 4809 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200050 4809 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200058 4809 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200067 4809 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200075 4809 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200082 4809 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200099 4809 flags.go:64] FLAG: --enable-load-reader="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200107 4809 flags.go:64] FLAG: --enable-server="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200115 4809 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200125 4809 flags.go:64] FLAG: --event-burst="100" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200133 4809 flags.go:64] FLAG: --event-qps="50" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200140 4809 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200147 4809 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200155 4809 flags.go:64] FLAG: --eviction-hard="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200178 4809 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200186 4809 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200194 4809 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200202 4809 flags.go:64] FLAG: --eviction-soft="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200209 4809 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200217 4809 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200224 4809 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200231 4809 flags.go:64] FLAG: --experimental-mounter-path="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200238 4809 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200245 4809 flags.go:64] FLAG: --fail-swap-on="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200252 4809 flags.go:64] FLAG: --feature-gates="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200263 4809 flags.go:64] FLAG: --file-check-frequency="20s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200271 4809 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200278 4809 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200286 4809 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200294 4809 flags.go:64] FLAG: --healthz-port="10248" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200301 4809 flags.go:64] FLAG: --help="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200309 4809 flags.go:64] FLAG: --hostname-override="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200317 4809 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200325 4809 flags.go:64] FLAG: --http-check-frequency="20s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200334 4809 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200343 4809 flags.go:64] FLAG: --image-credential-provider-config="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200351 4809 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200359 4809 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200367 4809 flags.go:64] FLAG: --image-service-endpoint="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200374 4809 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200381 4809 flags.go:64] FLAG: --kube-api-burst="100" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200389 4809 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200399 4809 flags.go:64] FLAG: --kube-api-qps="50" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200406 4809 flags.go:64] FLAG: --kube-reserved="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200413 4809 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200421 4809 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200432 4809 flags.go:64] FLAG: --kubelet-cgroups="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200439 4809 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200447 4809 flags.go:64] FLAG: --lock-file="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200453 4809 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200461 4809 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200468 4809 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200480 4809 flags.go:64] FLAG: --log-json-split-stream="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200487 4809 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200494 4809 flags.go:64] FLAG: --log-text-split-stream="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200501 4809 flags.go:64] FLAG: --logging-format="text" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200508 4809 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200516 4809 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200523 4809 flags.go:64] FLAG: --manifest-url="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200530 4809 flags.go:64] FLAG: --manifest-url-header="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200541 4809 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200548 4809 flags.go:64] FLAG: --max-open-files="1000000" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200558 4809 flags.go:64] FLAG: --max-pods="110" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200565 4809 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200571 4809 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200578 4809 flags.go:64] FLAG: --memory-manager-policy="None" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200586 4809 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200594 4809 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200602 4809 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200610 4809 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200630 4809 flags.go:64] FLAG: --node-status-max-images="50" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200637 4809 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200645 4809 flags.go:64] FLAG: --oom-score-adj="-999" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200652 4809 flags.go:64] FLAG: --pod-cidr="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200659 4809 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200672 4809 flags.go:64] FLAG: --pod-manifest-path="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200680 4809 flags.go:64] FLAG: --pod-max-pids="-1" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200687 4809 flags.go:64] FLAG: --pods-per-core="0" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200695 4809 flags.go:64] FLAG: --port="10250" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200704 4809 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200711 4809 flags.go:64] FLAG: --provider-id="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200718 4809 flags.go:64] FLAG: --qos-reserved="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200725 4809 flags.go:64] FLAG: --read-only-port="10255" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200732 4809 flags.go:64] FLAG: --register-node="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200765 4809 flags.go:64] FLAG: --register-schedulable="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200773 4809 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200787 4809 flags.go:64] FLAG: --registry-burst="10" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200794 4809 flags.go:64] FLAG: --registry-qps="5" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200802 4809 flags.go:64] FLAG: --reserved-cpus="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200808 4809 flags.go:64] FLAG: --reserved-memory="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200819 4809 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200826 4809 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200833 4809 flags.go:64] FLAG: --rotate-certificates="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200841 4809 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200848 4809 flags.go:64] FLAG: --runonce="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200856 4809 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200864 4809 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200873 4809 flags.go:64] FLAG: --seccomp-default="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200880 4809 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200887 4809 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200895 4809 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200902 4809 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200910 4809 flags.go:64] FLAG: --storage-driver-password="root" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200918 4809 flags.go:64] FLAG: --storage-driver-secure="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200925 4809 flags.go:64] FLAG: --storage-driver-table="stats" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200932 4809 flags.go:64] FLAG: --storage-driver-user="root" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200939 4809 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200947 4809 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200954 4809 flags.go:64] FLAG: --system-cgroups="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200962 4809 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200974 4809 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200981 4809 flags.go:64] FLAG: --tls-cert-file="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200989 4809 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.200998 4809 flags.go:64] FLAG: --tls-min-version="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201007 4809 flags.go:64] FLAG: --tls-private-key-file="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201014 4809 flags.go:64] FLAG: --topology-manager-policy="none" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201021 4809 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201028 4809 flags.go:64] FLAG: --topology-manager-scope="container" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201036 4809 flags.go:64] FLAG: --v="2" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201046 4809 flags.go:64] FLAG: --version="false" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201057 4809 flags.go:64] FLAG: --vmodule="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201067 4809 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201074 4809 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201307 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201319 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201328 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201336 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201344 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201350 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201356 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201363 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201371 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201379 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201386 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201392 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201400 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201444 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201452 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201461 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201468 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201477 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201484 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201491 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201500 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201508 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201514 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201522 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201528 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201534 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201542 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201552 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201560 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201582 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201590 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201597 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201604 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201612 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201619 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201626 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201635 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201642 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201650 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201656 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201662 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201668 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201674 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201681 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201687 4809 feature_gate.go:330] unrecognized feature gate: Example Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201693 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201698 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201705 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201711 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201718 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201723 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201729 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201761 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201768 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201774 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201781 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201786 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201793 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201799 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201807 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201814 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201825 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201832 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201840 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201846 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201853 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201867 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201874 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201881 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201888 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.201894 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.201913 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.219170 4809 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.219245 4809 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220132 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220178 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220191 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220201 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220212 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220222 4809 feature_gate.go:330] unrecognized feature gate: Example Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220232 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220241 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220249 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220257 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220266 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220276 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220291 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220315 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220329 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220340 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220350 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220364 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220373 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220382 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220391 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220405 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220421 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220433 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220444 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220456 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220466 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220475 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220484 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220493 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220503 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220523 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220532 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220542 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220553 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220565 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220574 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220584 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220593 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220602 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220611 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220619 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220627 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220636 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220645 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220654 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220662 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220670 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220681 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220692 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220701 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220710 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220720 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220729 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220789 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220808 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220821 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220832 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220841 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220851 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220861 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220870 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220878 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220887 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220896 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220907 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220918 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220928 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220937 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220945 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.220955 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.220970 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221214 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221226 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221236 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221247 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221259 4809 feature_gate.go:330] unrecognized feature gate: Example Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221268 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221277 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221286 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221294 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221302 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221313 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221324 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221332 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221342 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221350 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221359 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221367 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221378 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221387 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221396 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221450 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221461 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221470 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221479 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221487 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221496 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221505 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221513 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221522 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221531 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221539 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221548 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221556 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221564 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221573 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221581 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221589 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221601 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221609 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221618 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221626 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221635 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221643 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221651 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221660 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221668 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221676 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221684 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221692 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221703 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221711 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221720 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221728 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221773 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221785 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221801 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221815 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221828 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221839 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221850 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221859 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221869 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221878 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221886 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221895 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221907 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221918 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221927 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221936 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221944 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.221954 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.221969 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.222259 4809 server.go:940] "Client rotation is on, will bootstrap in background" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.229036 4809 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.229187 4809 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.231170 4809 server.go:997] "Starting client certificate rotation" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.231223 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.231432 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-07 11:04:47.001833576 +0000 UTC Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.231536 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.263615 4809 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.265941 4809 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.268420 4809 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.281777 4809 log.go:25] "Validated CRI v1 runtime API" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.314850 4809 log.go:25] "Validated CRI v1 image API" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.316795 4809 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.322407 4809 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-27-17-05-03-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.322463 4809 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.352729 4809 manager.go:217] Machine: {Timestamp:2025-11-27 17:09:25.347884672 +0000 UTC m=+0.620342064 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:2fa873ac-d151-4e97-8902-297a67778c76 BootID:37894acd-0695-4c66-9d28-ea598b1264ef Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8b:60:d3 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8b:60:d3 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:0a:77:58 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:38:70:de Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:e3:43:08 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a6:1d:cd Speed:-1 Mtu:1496} {Name:eth10 MacAddress:b2:61:47:23:57:56 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:e2:b5:ae:a6:02:e7 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.354472 4809 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.354749 4809 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.360716 4809 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.361451 4809 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.361495 4809 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.361759 4809 topology_manager.go:138] "Creating topology manager with none policy" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.361772 4809 container_manager_linux.go:303] "Creating device plugin manager" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.362221 4809 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.362255 4809 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.362465 4809 state_mem.go:36] "Initialized new in-memory state store" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.362588 4809 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.365802 4809 kubelet.go:418] "Attempting to sync node with API server" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.365827 4809 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.365852 4809 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.365866 4809 kubelet.go:324] "Adding apiserver pod source" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.365879 4809 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.371311 4809 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.373197 4809 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.373246 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.373270 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.373379 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.373395 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.375408 4809 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.376971 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377010 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377021 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377037 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377059 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377073 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377088 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377104 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377116 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377128 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377145 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.377157 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.381101 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.381726 4809 server.go:1280] "Started kubelet" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.382717 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:25 crc systemd[1]: Started Kubernetes Kubelet. Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.384411 4809 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.384952 4809 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.385061 4809 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.386402 4809 server.go:460] "Adding debug handlers to kubelet server" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.386570 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.386618 4809 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.386669 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 21:14:40.000203786 +0000 UTC Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.386725 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 652h5m14.61348122s for next certificate rotation Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.386803 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.386811 4809 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.386823 4809 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.386832 4809 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.387805 4809 factory.go:55] Registering systemd factory Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.387836 4809 factory.go:221] Registration of the systemd container factory successfully Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.387963 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.388803 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.388112 4809 factory.go:153] Registering CRI-O factory Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.388831 4809 factory.go:221] Registration of the crio container factory successfully Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.388922 4809 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.388945 4809 factory.go:103] Registering Raw factory Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.388960 4809 manager.go:1196] Started watching for new ooms in manager Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.389004 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="200ms" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.389490 4809 manager.go:319] Starting recovery of all containers Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.388800 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.248:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187bec2eefcf8d07 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 17:09:25.381688583 +0000 UTC m=+0.654145955,LastTimestamp:2025-11-27 17:09:25.381688583 +0000 UTC m=+0.654145955,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.407909 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408004 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408031 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408052 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408073 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408094 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408114 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408135 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408161 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408180 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408199 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408219 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408242 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408270 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408294 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408314 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408339 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408363 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408383 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408405 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408426 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408450 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408472 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408495 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408516 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408539 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408572 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408604 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408632 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408653 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408676 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408696 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408719 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408815 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408843 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408863 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408886 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408907 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408927 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408947 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408967 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.408991 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409014 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409034 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409053 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409079 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409100 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409120 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409147 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409168 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409189 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409208 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409236 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409260 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409280 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409347 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409374 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409397 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409420 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409452 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409478 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409499 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409521 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409542 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.409565 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410214 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410235 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410256 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410275 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410295 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410318 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410338 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410357 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410381 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410401 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410424 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410444 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410464 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410483 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410503 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410526 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410549 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410568 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410589 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410609 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410627 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410648 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410669 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410689 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410712 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410732 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410789 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410809 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.410828 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.411044 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.411064 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414503 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414551 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414573 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414592 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414609 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414626 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414648 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414663 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414694 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414716 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414732 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414763 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414777 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414792 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414807 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414820 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414836 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414849 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414862 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414873 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414887 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414898 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414909 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414923 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414936 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414948 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414961 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414973 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.414987 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415003 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415015 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415028 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415042 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415055 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415067 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415079 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415090 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415104 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415116 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415127 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415143 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415154 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415167 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415178 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415191 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415203 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415215 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415228 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415246 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415257 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415271 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415284 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415296 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415307 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415317 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415328 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415340 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415350 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415361 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415373 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415384 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415399 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415410 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415422 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415434 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415445 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415456 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415469 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415480 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415489 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415503 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415514 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415526 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415535 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415545 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415555 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415565 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415575 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415586 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415595 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.415606 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.416285 4809 manager.go:324] Recovery completed Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418053 4809 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418122 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418145 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418216 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418233 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418248 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418262 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418276 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418291 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418318 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418351 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418367 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418386 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418400 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418416 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418458 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418483 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418527 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418683 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418717 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418790 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418812 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418843 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418874 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418893 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418913 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418932 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418946 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.418991 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.419018 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.419032 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.419091 4809 reconstruct.go:97] "Volume reconstruction finished" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.419100 4809 reconciler.go:26] "Reconciler: start to sync state" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.431837 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.434097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.434157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.434167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.435034 4809 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.435051 4809 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.435074 4809 state_mem.go:36] "Initialized new in-memory state store" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.453652 4809 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.455507 4809 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.456438 4809 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.456549 4809 kubelet.go:2335] "Starting kubelet main sync loop" Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.456629 4809 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.459515 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.459668 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.459798 4809 policy_none.go:49] "None policy: Start" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.462410 4809 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.462447 4809 state_mem.go:35] "Initializing new in-memory state store" Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.487485 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.532113 4809 manager.go:334] "Starting Device Plugin manager" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.532454 4809 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.532527 4809 server.go:79] "Starting device plugin registration server" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.533183 4809 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.533299 4809 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.533528 4809 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.533790 4809 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.533822 4809 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.543577 4809 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.557834 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.558348 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.561275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.561334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.561344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.561562 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.563400 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.563493 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.567106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.567142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.567153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.567235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.567265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.567277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.567427 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.567824 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.567967 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.568255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.568300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.568311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.568537 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.568711 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.568769 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569451 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569556 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569791 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.569891 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.570210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.570261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.570272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.570430 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.570455 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.571556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.571576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.571584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.571681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.571700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.571710 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.590169 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="400ms" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621678 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621732 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621813 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621832 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621885 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621923 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621945 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621966 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621982 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.621997 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.622028 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.622045 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.622084 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.622119 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.633955 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.635301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.635367 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.635386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.635458 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.636124 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.248:6443: connect: connection refused" node="crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723609 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723636 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723656 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723676 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723761 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723786 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723762 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723807 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723785 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723823 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723837 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723841 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723852 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723810 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723871 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723880 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723866 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723897 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723904 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723890 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723956 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723656 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723924 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723915 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.724128 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723790 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.723914 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.836399 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.837784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.837840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.837862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.837901 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.838493 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.248:6443: connect: connection refused" node="crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.898703 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.906176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.926389 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.935030 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: I1127 17:09:25.939876 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.950396 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-2a7f3f06725dc7436724477a89874b1d95415574496309d0b178de800f1b131b WatchSource:0}: Error finding container 2a7f3f06725dc7436724477a89874b1d95415574496309d0b178de800f1b131b: Status 404 returned error can't find the container with id 2a7f3f06725dc7436724477a89874b1d95415574496309d0b178de800f1b131b Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.952025 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7dd05d0d2cd3a4139cda1501cb3622a7803f1bbc37e9e7c4c3c56e71bbabefc1 WatchSource:0}: Error finding container 7dd05d0d2cd3a4139cda1501cb3622a7803f1bbc37e9e7c4c3c56e71bbabefc1: Status 404 returned error can't find the container with id 7dd05d0d2cd3a4139cda1501cb3622a7803f1bbc37e9e7c4c3c56e71bbabefc1 Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.958095 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-f7c4fae74ba72bf93ae2991ef6bedf51c805d70d8acf72616c04e72eeaa68b94 WatchSource:0}: Error finding container f7c4fae74ba72bf93ae2991ef6bedf51c805d70d8acf72616c04e72eeaa68b94: Status 404 returned error can't find the container with id f7c4fae74ba72bf93ae2991ef6bedf51c805d70d8acf72616c04e72eeaa68b94 Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.968884 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-6651bd9c650e653b2e2081ca2554b970471877d112749de10934509f52939490 WatchSource:0}: Error finding container 6651bd9c650e653b2e2081ca2554b970471877d112749de10934509f52939490: Status 404 returned error can't find the container with id 6651bd9c650e653b2e2081ca2554b970471877d112749de10934509f52939490 Nov 27 17:09:25 crc kubenswrapper[4809]: W1127 17:09:25.972391 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ec65830445c2a0a3fedf706c760171c89ca4c6a06bc1d0e3e212d0a904ce6d54 WatchSource:0}: Error finding container ec65830445c2a0a3fedf706c760171c89ca4c6a06bc1d0e3e212d0a904ce6d54: Status 404 returned error can't find the container with id ec65830445c2a0a3fedf706c760171c89ca4c6a06bc1d0e3e212d0a904ce6d54 Nov 27 17:09:25 crc kubenswrapper[4809]: E1127 17:09:25.991637 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="800ms" Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.238836 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.240523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.240557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.240572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.240622 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 17:09:26 crc kubenswrapper[4809]: E1127 17:09:26.241344 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.248:6443: connect: connection refused" node="crc" Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.384520 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.464513 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7dd05d0d2cd3a4139cda1501cb3622a7803f1bbc37e9e7c4c3c56e71bbabefc1"} Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.465481 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2a7f3f06725dc7436724477a89874b1d95415574496309d0b178de800f1b131b"} Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.466420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ec65830445c2a0a3fedf706c760171c89ca4c6a06bc1d0e3e212d0a904ce6d54"} Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.467420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6651bd9c650e653b2e2081ca2554b970471877d112749de10934509f52939490"} Nov 27 17:09:26 crc kubenswrapper[4809]: I1127 17:09:26.468270 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f7c4fae74ba72bf93ae2991ef6bedf51c805d70d8acf72616c04e72eeaa68b94"} Nov 27 17:09:26 crc kubenswrapper[4809]: W1127 17:09:26.636895 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:26 crc kubenswrapper[4809]: E1127 17:09:26.637036 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:26 crc kubenswrapper[4809]: E1127 17:09:26.792982 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="1.6s" Nov 27 17:09:26 crc kubenswrapper[4809]: W1127 17:09:26.829482 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:26 crc kubenswrapper[4809]: E1127 17:09:26.829573 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:26 crc kubenswrapper[4809]: W1127 17:09:26.906374 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:26 crc kubenswrapper[4809]: E1127 17:09:26.906480 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:26 crc kubenswrapper[4809]: W1127 17:09:26.945197 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:26 crc kubenswrapper[4809]: E1127 17:09:26.945339 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.041603 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.043678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.043734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.043781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.043816 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 17:09:27 crc kubenswrapper[4809]: E1127 17:09:27.044369 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.248:6443: connect: connection refused" node="crc" Nov 27 17:09:27 crc kubenswrapper[4809]: E1127 17:09:27.115276 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.248:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187bec2eefcf8d07 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 17:09:25.381688583 +0000 UTC m=+0.654145955,LastTimestamp:2025-11-27 17:09:25.381688583 +0000 UTC m=+0.654145955,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.351974 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 27 17:09:27 crc kubenswrapper[4809]: E1127 17:09:27.354584 4809 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.384276 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.471993 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59" exitCode=0 Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.472081 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.472081 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59"} Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.473374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.473398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.473407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.474109 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5628dbf8173e6827b814898bfdd58d1df9e0ae22939e3756a3d1f71b8f7a123d" exitCode=0 Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.474182 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.474180 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5628dbf8173e6827b814898bfdd58d1df9e0ae22939e3756a3d1f71b8f7a123d"} Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.475095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.475119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.475130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.475873 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.476654 4809 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b" exitCode=0 Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.476725 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b"} Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.476843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.476864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.476873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.476903 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.478686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.478711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.478722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.479691 4809 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="725750eead54ee1b84c7ec6c15e16019f58e9bfdc099f28d286076b2907f8235" exitCode=0 Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.479841 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.479820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"725750eead54ee1b84c7ec6c15e16019f58e9bfdc099f28d286076b2907f8235"} Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.481407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.481437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.481450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.483968 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a"} Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.483998 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08"} Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.484009 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc"} Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.484019 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8"} Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.484143 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.485342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.485397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.485423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.840854 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:27 crc kubenswrapper[4809]: I1127 17:09:27.992335 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.384668 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:28 crc kubenswrapper[4809]: E1127 17:09:28.394475 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="3.2s" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.491711 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521"} Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.491790 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff"} Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.491809 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff"} Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.491940 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.493642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.493674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.493687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.496169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1be54422827bedc61e9f1a13630540aac14ef91a65d85dfea89dcc578e0d90c0"} Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.496241 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.497108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.497129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.497141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.500760 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798"} Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.500789 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd"} Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.500803 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda"} Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.500816 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e"} Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.502877 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="07ee0db804d52fa1651d5eee7bf0a1ddb81bf96e5a616cc5e7b915c149996901" exitCode=0 Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.502997 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.503564 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.504033 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"07ee0db804d52fa1651d5eee7bf0a1ddb81bf96e5a616cc5e7b915c149996901"} Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.504533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.504557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.504569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.505106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.505132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.505148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.645527 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.647079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.647125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.647138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:28 crc kubenswrapper[4809]: I1127 17:09:28.647276 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 17:09:28 crc kubenswrapper[4809]: E1127 17:09:28.648010 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.248:6443: connect: connection refused" node="crc" Nov 27 17:09:28 crc kubenswrapper[4809]: W1127 17:09:28.765688 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:28 crc kubenswrapper[4809]: E1127 17:09:28.765853 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:28 crc kubenswrapper[4809]: W1127 17:09:28.862538 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.248:6443: connect: connection refused Nov 27 17:09:28 crc kubenswrapper[4809]: E1127 17:09:28.862649 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.248:6443: connect: connection refused" logger="UnhandledError" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.507974 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f4824645026999e711ed4c5dc7f8857a33e0bfeac58cf6e77d798a9f9ab699a1" exitCode=0 Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.508096 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.508087 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f4824645026999e711ed4c5dc7f8857a33e0bfeac58cf6e77d798a9f9ab699a1"} Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.509070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.509107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.509119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.511791 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.511808 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.511824 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.511846 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.511982 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.512605 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"771b257f003ff5dedd7d726e0a812828ca215e4ce193efcce4411a9b3edba17f"} Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.513185 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.513290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.513301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.517675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.517714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.517755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.518251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.518391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.518511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.525980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.526005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:29 crc kubenswrapper[4809]: I1127 17:09:29.526014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.290178 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.517383 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a9c71189531579749b27b0c49e83d98646a4222aefc06e29b3f7822230dd25ab"} Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.517437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"34401cef59489038f8fb148eb3db776826ca6ce79a3428846d1b3f3d102e71ae"} Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.517443 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.517450 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"168aba809229b0e356f02151ec9fc48f2aafa1de882b9a080a17f36f38570c55"} Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.517541 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.517537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"847769dbc77952b4382055d8b6aacd5a6cfde6d3e5b5c6f3d2262698929c15d8"} Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.517597 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.518222 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.518245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.518254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.518436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.518465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.518478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.887806 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.993627 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 17:09:30 crc kubenswrapper[4809]: I1127 17:09:30.993803 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.149916 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.150179 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.152167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.152224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.152245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.471242 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.528620 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"61f28e0a9ca53636faae3bddfe1346a9bb416102145e38fff56ad1fd0d06f735"} Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.528687 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.528716 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.528807 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.530576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.530636 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.530643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.530660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.530684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.530709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.848943 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.850909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.850953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.850973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:31 crc kubenswrapper[4809]: I1127 17:09:31.851006 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.126353 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.126605 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.128371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.128404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.128413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.135920 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.530463 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.530525 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.531642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.531693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.531715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.531825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.531885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:32 crc kubenswrapper[4809]: I1127 17:09:32.531903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:33 crc kubenswrapper[4809]: I1127 17:09:33.819290 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 27 17:09:33 crc kubenswrapper[4809]: I1127 17:09:33.819511 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:33 crc kubenswrapper[4809]: I1127 17:09:33.821152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:33 crc kubenswrapper[4809]: I1127 17:09:33.821236 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:33 crc kubenswrapper[4809]: I1127 17:09:33.821262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:34 crc kubenswrapper[4809]: I1127 17:09:34.401425 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:34 crc kubenswrapper[4809]: I1127 17:09:34.401639 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 17:09:34 crc kubenswrapper[4809]: I1127 17:09:34.401682 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:34 crc kubenswrapper[4809]: I1127 17:09:34.402863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:34 crc kubenswrapper[4809]: I1127 17:09:34.402895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:34 crc kubenswrapper[4809]: I1127 17:09:34.402906 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:35 crc kubenswrapper[4809]: E1127 17:09:35.543768 4809 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.686885 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.687141 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.688478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.688527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.688543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.979612 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.979891 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.981368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.981417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:35 crc kubenswrapper[4809]: I1127 17:09:35.981433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.384203 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.554579 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.556312 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="771b257f003ff5dedd7d726e0a812828ca215e4ce193efcce4411a9b3edba17f" exitCode=255 Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.556360 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"771b257f003ff5dedd7d726e0a812828ca215e4ce193efcce4411a9b3edba17f"} Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.556527 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.557488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.557517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.557530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.558224 4809 scope.go:117] "RemoveContainer" containerID="771b257f003ff5dedd7d726e0a812828ca215e4ce193efcce4411a9b3edba17f" Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.645289 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.645369 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.657069 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 27 17:09:39 crc kubenswrapper[4809]: I1127 17:09:39.657345 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 27 17:09:40 crc kubenswrapper[4809]: I1127 17:09:40.562503 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 17:09:40 crc kubenswrapper[4809]: I1127 17:09:40.564592 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847"} Nov 27 17:09:40 crc kubenswrapper[4809]: I1127 17:09:40.564834 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:40 crc kubenswrapper[4809]: I1127 17:09:40.565840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:40 crc kubenswrapper[4809]: I1127 17:09:40.565868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:40 crc kubenswrapper[4809]: I1127 17:09:40.565877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:40 crc kubenswrapper[4809]: I1127 17:09:40.992629 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 17:09:40 crc kubenswrapper[4809]: I1127 17:09:40.992695 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 27 17:09:41 crc kubenswrapper[4809]: I1127 17:09:41.158484 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:41 crc kubenswrapper[4809]: I1127 17:09:41.158660 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:41 crc kubenswrapper[4809]: I1127 17:09:41.159861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:41 crc kubenswrapper[4809]: I1127 17:09:41.159904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:41 crc kubenswrapper[4809]: I1127 17:09:41.159913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:43 crc kubenswrapper[4809]: I1127 17:09:43.841444 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 27 17:09:43 crc kubenswrapper[4809]: I1127 17:09:43.841643 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:43 crc kubenswrapper[4809]: I1127 17:09:43.842906 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:43 crc kubenswrapper[4809]: I1127 17:09:43.842972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:43 crc kubenswrapper[4809]: I1127 17:09:43.842988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:43 crc kubenswrapper[4809]: I1127 17:09:43.853689 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.406683 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.406928 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.407023 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.408198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.408257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.408269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.412608 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.574367 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.574453 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.575358 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.575396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.575404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.575428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.575461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.575472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: E1127 17:09:44.643573 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.645613 4809 trace.go:236] Trace[460836704]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 17:09:29.777) (total time: 14867ms): Nov 27 17:09:44 crc kubenswrapper[4809]: Trace[460836704]: ---"Objects listed" error: 14867ms (17:09:44.645) Nov 27 17:09:44 crc kubenswrapper[4809]: Trace[460836704]: [14.867556593s] [14.867556593s] END Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.645653 4809 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.645631 4809 trace.go:236] Trace[1873410712]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 17:09:33.070) (total time: 11574ms): Nov 27 17:09:44 crc kubenswrapper[4809]: Trace[1873410712]: ---"Objects listed" error: 11574ms (17:09:44.645) Nov 27 17:09:44 crc kubenswrapper[4809]: Trace[1873410712]: [11.57479362s] [11.57479362s] END Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.645717 4809 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.649588 4809 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.649588 4809 trace.go:236] Trace[1452493498]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 17:09:29.687) (total time: 14962ms): Nov 27 17:09:44 crc kubenswrapper[4809]: Trace[1452493498]: ---"Objects listed" error: 14962ms (17:09:44.649) Nov 27 17:09:44 crc kubenswrapper[4809]: Trace[1452493498]: [14.962340902s] [14.962340902s] END Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.649646 4809 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.650056 4809 trace.go:236] Trace[774737275]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 17:09:34.141) (total time: 10508ms): Nov 27 17:09:44 crc kubenswrapper[4809]: Trace[774737275]: ---"Objects listed" error: 10508ms (17:09:44.649) Nov 27 17:09:44 crc kubenswrapper[4809]: Trace[774737275]: [10.508210475s] [10.508210475s] END Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.650095 4809 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.657102 4809 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.657423 4809 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.658920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.658964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.658977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.659000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.659014 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:44Z","lastTransitionTime":"2025-11-27T17:09:44Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.661698 4809 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.679485 4809 csr.go:261] certificate signing request csr-gzfcr is approved, waiting to be issued Nov 27 17:09:44 crc kubenswrapper[4809]: E1127 17:09:44.680989 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.686788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.686866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.686880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.686907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.686952 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:44Z","lastTransitionTime":"2025-11-27T17:09:44Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.696612 4809 csr.go:257] certificate signing request csr-gzfcr is issued Nov 27 17:09:44 crc kubenswrapper[4809]: E1127 17:09:44.701123 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.711051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.711086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.711094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.711113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.711122 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:44Z","lastTransitionTime":"2025-11-27T17:09:44Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:44 crc kubenswrapper[4809]: E1127 17:09:44.722935 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.732931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.732976 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.732989 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.733009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.733022 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:44Z","lastTransitionTime":"2025-11-27T17:09:44Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:44 crc kubenswrapper[4809]: E1127 17:09:44.753881 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.765555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.765600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.765608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.765631 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.765641 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:44Z","lastTransitionTime":"2025-11-27T17:09:44Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:44 crc kubenswrapper[4809]: E1127 17:09:44.779657 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:44Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:44 crc kubenswrapper[4809]: E1127 17:09:44.779852 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.781625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.781722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.781808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.781874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.781938 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:44Z","lastTransitionTime":"2025-11-27T17:09:44Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.883858 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.883893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.883902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.883918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.883929 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:44Z","lastTransitionTime":"2025-11-27T17:09:44Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.986367 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.986465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.986492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.986534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:44 crc kubenswrapper[4809]: I1127 17:09:44.986560 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:44Z","lastTransitionTime":"2025-11-27T17:09:44Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.089926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.089977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.089990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.090058 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.090077 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:45Z","lastTransitionTime":"2025-11-27T17:09:45Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.192137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.192182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.192194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.192212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.192229 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:45Z","lastTransitionTime":"2025-11-27T17:09:45Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.231638 4809 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 27 17:09:45 crc kubenswrapper[4809]: W1127 17:09:45.231929 4809 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Node ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Nov 27 17:09:45 crc kubenswrapper[4809]: W1127 17:09:45.231960 4809 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Service ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Nov 27 17:09:45 crc kubenswrapper[4809]: W1127 17:09:45.231960 4809 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Nov 27 17:09:45 crc kubenswrapper[4809]: W1127 17:09:45.231978 4809 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.294516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.294552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.294560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.294578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.294588 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:45Z","lastTransitionTime":"2025-11-27T17:09:45Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.378451 4809 apiserver.go:52] "Watching apiserver" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.391222 4809 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.391887 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-qwx9w","openshift-multus/multus-additional-cni-plugins-s4g6j","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-ovn-kubernetes/ovnkube-node-m5b98","openshift-image-registry/node-ca-fr7qb","openshift-multus/multus-gd8px","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-d6cbb"] Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.392857 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.392928 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393150 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393205 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393357 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-d6cbb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393410 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393440 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393460 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393485 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393527 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393635 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.393705 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.394056 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.394143 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.394185 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.396204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.396239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.396251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.396268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.396281 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:45Z","lastTransitionTime":"2025-11-27T17:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.402050 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.402085 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.402134 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.402194 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.402467 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.402473 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.402599 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.402763 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403010 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403396 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403100 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403419 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403121 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403469 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403549 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403633 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403695 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403727 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403816 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403882 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403918 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403927 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403990 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.404001 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.404011 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.404113 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.404238 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.404270 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.404343 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403423 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403135 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403144 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403150 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403158 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.403125 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.427487 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.439785 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.449572 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.469952 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.480630 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.488519 4809 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.493789 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.498486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.498526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.498536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.498555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.498567 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:45Z","lastTransitionTime":"2025-11-27T17:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.507167 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.524529 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.554318 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.554381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.554428 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.554450 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.554475 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.554500 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.554524 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.554550 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.554992 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555006 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555002 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555067 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555115 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555138 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555159 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555177 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555259 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555407 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555461 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555686 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555692 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555674 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555952 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.556070 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.564800 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.555195 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.580987 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581011 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581050 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581070 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581090 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581108 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581123 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581146 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581162 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581178 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581196 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581210 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581224 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581241 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581261 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581276 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581293 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581309 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581353 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581389 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581408 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581426 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581441 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581460 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581481 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581500 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581518 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581536 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581555 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581574 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581730 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581792 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581820 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581850 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581873 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581911 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581930 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581947 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581964 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.581983 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582011 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582029 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582048 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582064 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582082 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582100 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582117 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582140 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582165 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582184 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582204 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582222 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582258 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582274 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582309 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582326 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582361 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582379 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582396 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582414 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582432 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582469 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582474 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582508 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582546 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582567 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582588 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582609 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582627 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582645 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582661 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582679 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582696 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582714 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582733 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.582767 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.583939 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.584185 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.584443 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.584456 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.584873 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585217 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585310 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585468 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585640 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585710 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585791 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585812 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585897 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585997 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.585831 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586142 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586178 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586201 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586223 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586268 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586307 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586328 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586346 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586364 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586399 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586417 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586438 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586454 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586471 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586496 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586514 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586959 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.587005 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.587024 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.587042 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.587917 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.587976 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.587994 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588017 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588039 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588098 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588127 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588145 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588165 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588184 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588203 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588229 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588244 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588260 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588611 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588640 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588657 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588676 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588693 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590544 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590576 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590595 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590614 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590801 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590850 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590868 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590889 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590919 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590938 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590955 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590972 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591007 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591026 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591043 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591060 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591078 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591114 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591156 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591174 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591194 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591211 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591227 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591245 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591261 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591279 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591294 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591311 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591328 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591345 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591361 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591377 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591394 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591411 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591445 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591463 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591479 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591494 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591512 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591529 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591545 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591561 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591576 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591593 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591609 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591625 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591652 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586391 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586545 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586653 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586366 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586950 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.586982 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.587103 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.587224 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.587556 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588460 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588633 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588707 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588765 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.588867 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.589101 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.589327 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.589431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.589436 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.589583 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.589829 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.589888 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590041 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590253 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590464 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590557 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590555 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.590655 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591055 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591084 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591372 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591384 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591591 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.591597 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605185 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605311 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605343 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605366 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605389 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605416 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605436 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605458 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605481 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605579 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-system-cni-dir\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605607 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-cni-dir\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.605625 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-daemon-config\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.609393 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-slash\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.609601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-script-lib\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.609675 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b-hosts-file\") pod \"node-resolver-d6cbb\" (UID: \"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\") " pod="openshift-dns/node-resolver-d6cbb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.609753 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.609805 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdbfe7ca-6631-4c54-a71a-4b73cae5a64a-host\") pod \"node-ca-fr7qb\" (UID: \"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\") " pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.609922 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.609949 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.609978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.610103 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-conf-dir\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.610130 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/21848912-a128-4704-b84a-811af21e99cc-cni-binary-copy\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.610278 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-systemd\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.610307 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-proxy-tls\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.611137 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.611135 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.611227 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.611245 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.611471 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.611525 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.611806 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.611885 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.611902 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.612055 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.612112 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.612278 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.612374 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.612656 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.612974 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.613555 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.614270 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.614532 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.614622 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.615033 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.616033 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.616539 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.617395 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.617464 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.617732 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.617805 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.618054 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.618222 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.618370 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.620027 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.620347 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.620962 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.621030 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.621810 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.622109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.622660 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.623043 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.623538 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.623777 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.623935 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.623995 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.624303 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.624350 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.624664 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.625142 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.625629 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.625840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.626141 4809 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.626189 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.626378 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.626791 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8741a6e8-a7f2-40ef-b25b-10e518345478-cni-binary-copy\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.626868 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.627091 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.627297 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.628819 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.629175 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.629569 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.629770 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.629866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.630211 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.630520 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.630707 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.634839 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.635158 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.635286 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.635590 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.635873 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.636032 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.636378 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.636503 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.636893 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.637132 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.637175 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.637337 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.637530 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-cnibin\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.637994 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.640115 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.637596 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-var-lib-cni-multus\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.640876 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.641418 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.644862 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.657073 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.657285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.657902 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.658423 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.658421 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.658555 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.658632 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.658882 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.659154 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.659615 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.660030 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.660336 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.662283 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.663065 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.664709 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.667102 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.667165 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.667611 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.667825 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.668176 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.668235 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.668391 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.668654 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.668725 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.668932 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.669764 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.671867 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.671920 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.671856 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.672347 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.672348 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.672628 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.672700 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.673352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.673855 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.674392 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.674394 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.674879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.674918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.674943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.674962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.674976 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.674977 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:45Z","lastTransitionTime":"2025-11-27T17:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.675089 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.675805 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.675815 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:46.175784113 +0000 UTC m=+21.448241465 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.676245 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:09:46.176234275 +0000 UTC m=+21.448691627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.676492 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.676640 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-529wf\" (UniqueName: \"kubernetes.io/projected/8741a6e8-a7f2-40ef-b25b-10e518345478-kube-api-access-529wf\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.676684 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.676819 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.677168 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.676784 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-bin\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.677626 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-run-netns\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.678462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-var-lib-cni-bin\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.678501 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.678910 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.680541 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-openvswitch\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.680892 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-var-lib-kubelet\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.682448 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.682560 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.683124 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.683528 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.683578 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.685048 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.685448 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.686063 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.686207 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689416 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-os-release\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689472 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-config\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689515 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689508 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8bqn\" (UniqueName: \"kubernetes.io/projected/cdbfe7ca-6631-4c54-a71a-4b73cae5a64a-kube-api-access-d8bqn\") pod \"node-ca-fr7qb\" (UID: \"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\") " pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689641 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/21848912-a128-4704-b84a-811af21e99cc-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689712 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689754 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689774 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-hostroot\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689808 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-run-multus-certs\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689827 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-tuning-conf-dir\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689844 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-netns\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.689926 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690038 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690061 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690081 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-systemd-units\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690099 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-ovn-kubernetes\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690117 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-cnibin\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690133 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-socket-dir-parent\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690155 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdhtq\" (UniqueName: \"kubernetes.io/projected/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-kube-api-access-pdhtq\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690172 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-etc-kubernetes\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690191 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-system-cni-dir\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-var-lib-openvswitch\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690272 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-node-log\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690292 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-env-overrides\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-run-k8s-cni-cncf-io\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690331 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2jxd\" (UniqueName: \"kubernetes.io/projected/21848912-a128-4704-b84a-811af21e99cc-kube-api-access-n2jxd\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690370 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-kubelet\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.690938 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.690989 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.690999 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:46.190982848 +0000 UTC m=+21.463440200 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691061 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgzf6\" (UniqueName: \"kubernetes.io/projected/42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b-kube-api-access-hgzf6\") pod \"node-resolver-d6cbb\" (UID: \"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\") " pod="openshift-dns/node-resolver-d6cbb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691095 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691128 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-mcd-auth-proxy-config\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691188 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-os-release\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691260 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-log-socket\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66b4318a-f089-451d-8a16-97de26acce28-ovn-node-metrics-cert\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691345 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-netd\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691371 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52b7c\" (UniqueName: \"kubernetes.io/projected/66b4318a-f089-451d-8a16-97de26acce28-kube-api-access-52b7c\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691395 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cdbfe7ca-6631-4c54-a71a-4b73cae5a64a-serviceca\") pod \"node-ca-fr7qb\" (UID: \"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\") " pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691435 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-etc-openvswitch\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691460 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-ovn\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691484 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-rootfs\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691692 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691718 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691731 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691763 4809 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691777 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691789 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691803 4809 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691816 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691828 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691840 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691851 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691865 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691878 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691891 4809 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691905 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691929 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691943 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691958 4809 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691970 4809 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691980 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.691992 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692005 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692016 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692030 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692044 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692058 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692069 4809 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692081 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692093 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692105 4809 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692117 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692129 4809 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692141 4809 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692154 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692169 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692182 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692197 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692210 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692224 4809 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692236 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692250 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692271 4809 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692284 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692298 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692310 4809 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692338 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692354 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692369 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692384 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692398 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692411 4809 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692424 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692436 4809 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692449 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692461 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692473 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692486 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692498 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692510 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692522 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692534 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692546 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692558 4809 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692569 4809 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692580 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692591 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692603 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692615 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692629 4809 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692640 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692651 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692662 4809 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692672 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692683 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692693 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692705 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692717 4809 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692729 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692758 4809 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692771 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692797 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692808 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692819 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692831 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692847 4809 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692859 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692870 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692881 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692892 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692908 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692919 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692930 4809 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692940 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692953 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692965 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.692996 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693036 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693050 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693061 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693073 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693087 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693102 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693121 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693135 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693147 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693159 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693171 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693184 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693195 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693205 4809 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693218 4809 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693229 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693241 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693252 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693263 4809 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693277 4809 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693289 4809 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693302 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693315 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693327 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693339 4809 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693351 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693363 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693375 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693387 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693398 4809 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693410 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693412 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693422 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693491 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693506 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693518 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693530 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693542 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693553 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693564 4809 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693574 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693586 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693588 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693609 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693600 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693681 4809 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693773 4809 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693823 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693831 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693862 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693905 4809 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.693926 4809 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.696046 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.696706 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.696721 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.698196 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-27 17:04:44 +0000 UTC, rotation deadline is 2026-10-09 08:08:21.360077213 +0000 UTC Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.698306 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7574h58m35.661776025s for next certificate rotation Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.700433 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701433 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701575 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701593 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701604 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701621 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701724 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701755 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701768 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701778 4809 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701788 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701798 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701811 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701918 4809 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701932 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701942 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701954 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.701964 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702079 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702098 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702113 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702127 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702245 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702262 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702275 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702287 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702296 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702409 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702420 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702431 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702442 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702452 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702567 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702582 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702594 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702604 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.702702 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.706335 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.706464 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.706480 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.706490 4809 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.706504 4809 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.706957 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.711923 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.711955 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.711968 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.712069 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:46.212028473 +0000 UTC m=+21.484485825 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.713569 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.715396 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.715802 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.722213 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.722264 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.722282 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:45 crc kubenswrapper[4809]: E1127 17:09:45.722358 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:46.222333954 +0000 UTC m=+21.494791306 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.725215 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.729665 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.745347 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.758767 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.771200 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.778531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.778581 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.778593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.778613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.778626 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:45Z","lastTransitionTime":"2025-11-27T17:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.782708 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771b257f003ff5dedd7d726e0a812828ca215e4ce193efcce4411a9b3edba17f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:39Z\\\",\\\"message\\\":\\\"W1127 17:09:28.681678 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:28.682477 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263368 cert, and key in /tmp/serving-cert-3585230431/serving-signer.crt, /tmp/serving-cert-3585230431/serving-signer.key\\\\nI1127 17:09:28.958087 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:28.960667 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 17:09:28.960847 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:28.961597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3585230431/tls.crt::/tmp/serving-cert-3585230431/tls.key\\\\\\\"\\\\nF1127 17:09:39.334091 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.800906 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807222 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/21848912-a128-4704-b84a-811af21e99cc-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807257 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-netns\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807288 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-hostroot\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807307 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-run-multus-certs\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807329 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-tuning-conf-dir\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807364 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-cnibin\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807384 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-socket-dir-parent\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-systemd-units\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-ovn-kubernetes\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807440 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-node-log\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807457 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-env-overrides\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807474 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdhtq\" (UniqueName: \"kubernetes.io/projected/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-kube-api-access-pdhtq\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807494 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-etc-kubernetes\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807511 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-system-cni-dir\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-var-lib-openvswitch\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807556 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgzf6\" (UniqueName: \"kubernetes.io/projected/42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b-kube-api-access-hgzf6\") pod \"node-resolver-d6cbb\" (UID: \"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\") " pod="openshift-dns/node-resolver-d6cbb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-run-k8s-cni-cncf-io\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807594 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2jxd\" (UniqueName: \"kubernetes.io/projected/21848912-a128-4704-b84a-811af21e99cc-kube-api-access-n2jxd\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807610 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-kubelet\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807632 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807659 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-mcd-auth-proxy-config\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807677 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66b4318a-f089-451d-8a16-97de26acce28-ovn-node-metrics-cert\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807712 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-os-release\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807732 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807769 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-log-socket\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807786 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cdbfe7ca-6631-4c54-a71a-4b73cae5a64a-serviceca\") pod \"node-ca-fr7qb\" (UID: \"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\") " pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807804 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-netd\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807870 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52b7c\" (UniqueName: \"kubernetes.io/projected/66b4318a-f089-451d-8a16-97de26acce28-kube-api-access-52b7c\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807890 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-etc-openvswitch\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807912 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-ovn\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807933 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-rootfs\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-etc-kubernetes\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807992 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-cnibin\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808033 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-system-cni-dir\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-hostroot\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.807961 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdbfe7ca-6631-4c54-a71a-4b73cae5a64a-host\") pod \"node-ca-fr7qb\" (UID: \"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\") " pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-var-lib-openvswitch\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808096 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-run-multus-certs\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808127 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-system-cni-dir\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808151 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-cni-dir\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808174 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-daemon-config\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808195 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-slash\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808213 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-script-lib\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808231 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b-hosts-file\") pod \"node-resolver-d6cbb\" (UID: \"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\") " pod="openshift-dns/node-resolver-d6cbb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808252 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-conf-dir\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808275 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8741a6e8-a7f2-40ef-b25b-10e518345478-cni-binary-copy\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/21848912-a128-4704-b84a-811af21e99cc-cni-binary-copy\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808276 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-socket-dir-parent\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808319 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-run-k8s-cni-cncf-io\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-systemd\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808359 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-proxy-tls\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808386 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-slash\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808380 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-var-lib-cni-multus\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-529wf\" (UniqueName: \"kubernetes.io/projected/8741a6e8-a7f2-40ef-b25b-10e518345478-kube-api-access-529wf\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808460 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-cnibin\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808479 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-openvswitch\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808495 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-bin\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808514 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-run-netns\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808531 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-var-lib-cni-bin\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808549 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-var-lib-kubelet\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808567 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8bqn\" (UniqueName: \"kubernetes.io/projected/cdbfe7ca-6631-4c54-a71a-4b73cae5a64a-kube-api-access-d8bqn\") pod \"node-ca-fr7qb\" (UID: \"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\") " pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808583 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-os-release\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808599 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-config\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808634 4809 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808646 4809 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808657 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808667 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808694 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-tuning-conf-dir\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808343 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-systemd\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.808010 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdbfe7ca-6631-4c54-a71a-4b73cae5a64a-host\") pod \"node-ca-fr7qb\" (UID: \"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\") " pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809020 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809092 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-os-release\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809301 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-script-lib\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809342 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-daemon-config\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/21848912-a128-4704-b84a-811af21e99cc-cnibin\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809466 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-openvswitch\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809523 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-os-release\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809556 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-bin\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809590 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-run-netns\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809624 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/21848912-a128-4704-b84a-811af21e99cc-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809631 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809667 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-var-lib-cni-bin\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809685 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-var-lib-kubelet\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809700 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-log-socket\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809710 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809763 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-kubelet\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809803 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-netns\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809833 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-node-log\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809859 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-netd\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-systemd-units\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.809921 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-ovn-kubernetes\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810002 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-config\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810047 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-host-var-lib-cni-multus\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-etc-openvswitch\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810097 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-ovn\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-rootfs\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810319 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-mcd-auth-proxy-config\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810332 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cdbfe7ca-6631-4c54-a71a-4b73cae5a64a-serviceca\") pod \"node-ca-fr7qb\" (UID: \"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\") " pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810400 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-cni-dir\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810445 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-system-cni-dir\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810469 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8741a6e8-a7f2-40ef-b25b-10e518345478-multus-conf-dir\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810509 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b-hosts-file\") pod \"node-resolver-d6cbb\" (UID: \"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\") " pod="openshift-dns/node-resolver-d6cbb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810529 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810548 4809 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810564 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810838 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-env-overrides\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810937 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/21848912-a128-4704-b84a-811af21e99cc-cni-binary-copy\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.810944 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8741a6e8-a7f2-40ef-b25b-10e518345478-cni-binary-copy\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.811370 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.813460 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66b4318a-f089-451d-8a16-97de26acce28-ovn-node-metrics-cert\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.813652 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-proxy-tls\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.819233 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.825823 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8bqn\" (UniqueName: \"kubernetes.io/projected/cdbfe7ca-6631-4c54-a71a-4b73cae5a64a-kube-api-access-d8bqn\") pod \"node-ca-fr7qb\" (UID: \"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\") " pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.826730 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgzf6\" (UniqueName: \"kubernetes.io/projected/42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b-kube-api-access-hgzf6\") pod \"node-resolver-d6cbb\" (UID: \"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\") " pod="openshift-dns/node-resolver-d6cbb" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.827233 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-529wf\" (UniqueName: \"kubernetes.io/projected/8741a6e8-a7f2-40ef-b25b-10e518345478-kube-api-access-529wf\") pod \"multus-gd8px\" (UID: \"8741a6e8-a7f2-40ef-b25b-10e518345478\") " pod="openshift-multus/multus-gd8px" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.827261 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2jxd\" (UniqueName: \"kubernetes.io/projected/21848912-a128-4704-b84a-811af21e99cc-kube-api-access-n2jxd\") pod \"multus-additional-cni-plugins-s4g6j\" (UID: \"21848912-a128-4704-b84a-811af21e99cc\") " pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.829474 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdhtq\" (UniqueName: \"kubernetes.io/projected/6b63ded2-aa4f-4aba-b3cc-a3965a01036c-kube-api-access-pdhtq\") pod \"machine-config-daemon-qwx9w\" (UID: \"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\") " pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.829901 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52b7c\" (UniqueName: \"kubernetes.io/projected/66b4318a-f089-451d-8a16-97de26acce28-kube-api-access-52b7c\") pod \"ovnkube-node-m5b98\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.830815 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.839997 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.852646 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.863421 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.874207 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.880773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.880820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.880832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.880850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.880863 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:45Z","lastTransitionTime":"2025-11-27T17:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.887376 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.899569 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.908822 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.919310 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.983820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.983865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.983876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.983893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:45 crc kubenswrapper[4809]: I1127 17:09:45.983950 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:45Z","lastTransitionTime":"2025-11-27T17:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.009170 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.020418 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-d6cbb" Nov 27 17:09:46 crc kubenswrapper[4809]: W1127 17:09:46.021567 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-f4bcf81e8aeeea163a307e71d3a208eee81bd2c3996f782cacb9c988e858a4f0 WatchSource:0}: Error finding container f4bcf81e8aeeea163a307e71d3a208eee81bd2c3996f782cacb9c988e858a4f0: Status 404 returned error can't find the container with id f4bcf81e8aeeea163a307e71d3a208eee81bd2c3996f782cacb9c988e858a4f0 Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.030941 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.041867 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.045666 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fr7qb" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.053052 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.067982 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gd8px" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.074252 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.081798 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.086936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.086969 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.086980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.086998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.087011 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:46Z","lastTransitionTime":"2025-11-27T17:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: W1127 17:09:46.123934 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8741a6e8_a7f2_40ef_b25b_10e518345478.slice/crio-2b0cbe5c0b84a8b832d3b66ed112823c0eb264d049e197e3b2719ad54307b15d WatchSource:0}: Error finding container 2b0cbe5c0b84a8b832d3b66ed112823c0eb264d049e197e3b2719ad54307b15d: Status 404 returned error can't find the container with id 2b0cbe5c0b84a8b832d3b66ed112823c0eb264d049e197e3b2719ad54307b15d Nov 27 17:09:46 crc kubenswrapper[4809]: W1127 17:09:46.127192 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-3e160052108045f5a4a55483d0d9e5815eeb9be7f6e4952801f4f4fdb7dcca2c WatchSource:0}: Error finding container 3e160052108045f5a4a55483d0d9e5815eeb9be7f6e4952801f4f4fdb7dcca2c: Status 404 returned error can't find the container with id 3e160052108045f5a4a55483d0d9e5815eeb9be7f6e4952801f4f4fdb7dcca2c Nov 27 17:09:46 crc kubenswrapper[4809]: W1127 17:09:46.128439 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-e59bdec6f63612d95bebd6e4c69e43e43020e1164e6aa67a69647e2757bcee47 WatchSource:0}: Error finding container e59bdec6f63612d95bebd6e4c69e43e43020e1164e6aa67a69647e2757bcee47: Status 404 returned error can't find the container with id e59bdec6f63612d95bebd6e4c69e43e43020e1164e6aa67a69647e2757bcee47 Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.193252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.193286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.193297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.193314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.193326 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:46Z","lastTransitionTime":"2025-11-27T17:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.216442 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.216580 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.216629 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.216675 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.216713 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.216757 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:09:47.216676538 +0000 UTC m=+22.489133890 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.216803 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:47.216782151 +0000 UTC m=+22.489239553 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.216819 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.216835 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.216853 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.216897 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:47.216880944 +0000 UTC m=+22.489338296 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.216974 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.217003 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:47.216995537 +0000 UTC m=+22.489452889 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.296522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.296568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.296580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.296600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.296612 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:46Z","lastTransitionTime":"2025-11-27T17:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.317336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.317895 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.317918 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.317930 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.317983 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:47.317969115 +0000 UTC m=+22.590426467 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.399428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.399464 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.399473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.399670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.399679 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:46Z","lastTransitionTime":"2025-11-27T17:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.502038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.502075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.502085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.502108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.502130 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:46Z","lastTransitionTime":"2025-11-27T17:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.588024 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.588071 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.588084 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f4bcf81e8aeeea163a307e71d3a208eee81bd2c3996f782cacb9c988e858a4f0"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.589796 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f" exitCode=0 Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.589876 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.589961 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"f5e01e2601c4021d050c567e3eaa7fbaff2144102276775b24d497d2ba951052"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.593222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.593271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.593282 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"2a49fad37886c13e993b9e013e186c3c3872b175ac3ff4ef11d3a94ed3a6de7d"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.595829 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.596597 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.598689 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847" exitCode=255 Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.598787 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.598881 4809 scope.go:117] "RemoveContainer" containerID="771b257f003ff5dedd7d726e0a812828ca215e4ce193efcce4411a9b3edba17f" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.599359 4809 scope.go:117] "RemoveContainer" containerID="13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847" Nov 27 17:09:46 crc kubenswrapper[4809]: E1127 17:09:46.599559 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.601338 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.601401 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e59bdec6f63612d95bebd6e4c69e43e43020e1164e6aa67a69647e2757bcee47"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.604650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.604704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.604719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.604777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.604794 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:46Z","lastTransitionTime":"2025-11-27T17:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.604830 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fr7qb" event={"ID":"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a","Type":"ContainerStarted","Data":"93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.604893 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fr7qb" event={"ID":"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a","Type":"ContainerStarted","Data":"179daebc347c515eee6d52f54b1da5a3222ed4aa6d3bb59f3f3f39dda74d8082"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.606671 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-d6cbb" event={"ID":"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b","Type":"ContainerStarted","Data":"e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.606759 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-d6cbb" event={"ID":"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b","Type":"ContainerStarted","Data":"928b62df97e5f2f71f939f61842165ff805e2b2a5ba8c11e7755fb4fce5ea274"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.609516 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.613555 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd8px" event={"ID":"8741a6e8-a7f2-40ef-b25b-10e518345478","Type":"ContainerStarted","Data":"e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.613594 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd8px" event={"ID":"8741a6e8-a7f2-40ef-b25b-10e518345478","Type":"ContainerStarted","Data":"2b0cbe5c0b84a8b832d3b66ed112823c0eb264d049e197e3b2719ad54307b15d"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.616492 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerStarted","Data":"209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.616530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerStarted","Data":"1ef1cfb0eabbdf7dd413087807b66f9b19d4dc0230088d335a01d1277fac84cf"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.618927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3e160052108045f5a4a55483d0d9e5815eeb9be7f6e4952801f4f4fdb7dcca2c"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.621372 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.631210 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.649042 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.659654 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.673146 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771b257f003ff5dedd7d726e0a812828ca215e4ce193efcce4411a9b3edba17f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:39Z\\\",\\\"message\\\":\\\"W1127 17:09:28.681678 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:28.682477 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263368 cert, and key in /tmp/serving-cert-3585230431/serving-signer.crt, /tmp/serving-cert-3585230431/serving-signer.key\\\\nI1127 17:09:28.958087 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:28.960667 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 17:09:28.960847 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:28.961597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3585230431/tls.crt::/tmp/serving-cert-3585230431/tls.key\\\\\\\"\\\\nF1127 17:09:39.334091 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.683245 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.695004 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.710062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.710115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.710144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.710169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.710106 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.710182 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:46Z","lastTransitionTime":"2025-11-27T17:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.725245 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.739153 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.755011 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.767284 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.781455 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.794658 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.809574 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.812664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.812706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.812718 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.812758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.812771 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:46Z","lastTransitionTime":"2025-11-27T17:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.822936 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.833929 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.846322 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.857102 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.870076 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.888705 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771b257f003ff5dedd7d726e0a812828ca215e4ce193efcce4411a9b3edba17f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:39Z\\\",\\\"message\\\":\\\"W1127 17:09:28.681678 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:28.682477 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263368 cert, and key in /tmp/serving-cert-3585230431/serving-signer.crt, /tmp/serving-cert-3585230431/serving-signer.key\\\\nI1127 17:09:28.958087 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:28.960667 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 17:09:28.960847 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:28.961597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3585230431/tls.crt::/tmp/serving-cert-3585230431/tls.key\\\\\\\"\\\\nF1127 17:09:39.334091 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.912050 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.916187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.916229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.916237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.916255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.916265 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:46Z","lastTransitionTime":"2025-11-27T17:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.924733 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.942276 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:46 crc kubenswrapper[4809]: I1127 17:09:46.955533 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:46Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.018997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.019031 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.019040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.019054 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.019064 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.120925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.120961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.120971 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.120986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.120995 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.222986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.223026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.223035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.223050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.223061 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.226854 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.226949 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.226983 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.227007 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.227075 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:09:49.227032816 +0000 UTC m=+24.499490168 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.227118 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.227146 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.227170 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.227183 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.227190 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:49.227171521 +0000 UTC m=+24.499628873 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.227189 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.227231 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:49.227217102 +0000 UTC m=+24.499674454 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.227363 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:49.227325605 +0000 UTC m=+24.499782947 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.325561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.325624 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.325649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.325677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.325690 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.328463 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.328677 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.328707 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.328721 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.328799 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:49.328781486 +0000 UTC m=+24.601238838 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.428224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.428269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.428280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.428301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.428336 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.457795 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.457807 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.457932 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.457970 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.458066 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.458160 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.461553 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.462524 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.463720 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.464386 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.465682 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.466297 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.466920 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.467911 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.468698 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.469773 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.470281 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.471314 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.471902 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.472478 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.473412 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.474010 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.475056 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.475416 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.476098 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.477147 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.477658 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.478814 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.479490 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.480838 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.481431 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.482395 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.483638 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.484226 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.485634 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.486251 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.487474 4809 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.487579 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.490763 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.491327 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.491779 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.492996 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.494027 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.494600 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.495265 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.496020 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.496584 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.497284 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.498022 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.498675 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.499233 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.503076 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.504280 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.505769 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.506306 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.507015 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.507965 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.508725 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.510107 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.510585 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.532490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.532542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.532553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.532570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.532580 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.627476 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.627561 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.627582 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.627603 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.627622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.627642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.629321 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.632793 4809 scope.go:117] "RemoveContainer" containerID="13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847" Nov 27 17:09:47 crc kubenswrapper[4809]: E1127 17:09:47.633514 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.635349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.635384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.635397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.635413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.635427 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.636405 4809 generic.go:334] "Generic (PLEG): container finished" podID="21848912-a128-4704-b84a-811af21e99cc" containerID="209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5" exitCode=0 Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.636566 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerDied","Data":"209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.656398 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.672554 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.688494 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.704520 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.721422 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.734958 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.738499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.738526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.738535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.738549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.738560 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.749909 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.766956 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.783690 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.802401 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.821445 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.835586 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.840421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.840449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.840458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.840473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.840482 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.857007 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.875716 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.899372 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.913103 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.923855 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.937546 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.942510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.942553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.942563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.942577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.942587 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:47Z","lastTransitionTime":"2025-11-27T17:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.951639 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.968473 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.982097 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:47 crc kubenswrapper[4809]: I1127 17:09:47.996266 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.000291 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.000329 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:47Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.005590 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.013304 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.026855 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.036239 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.044942 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.044978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.044987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.045002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.045013 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.052506 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.070116 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.090971 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.109007 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.121488 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.135037 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.147641 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.147666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.147675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.147689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.147700 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.148241 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.163169 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.176582 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.189784 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.204382 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.218641 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.238442 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.249539 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.249862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.249880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.249888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.249902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.249910 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.264916 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.352003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.352042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.352054 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.352073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.352083 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.454165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.454209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.454218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.454234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.454243 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.556425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.556464 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.556477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.556494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.556505 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.641065 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerStarted","Data":"5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.656805 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.658988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.659034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.659048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.659070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.659086 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.672233 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.686244 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.698452 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.710524 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.723237 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.737131 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.758660 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.761677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.761714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.761727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.761762 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.761776 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.779498 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.799359 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.813763 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.822432 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.833528 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.848841 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:48Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.864895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.864950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.864964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.864988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.865038 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.967931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.967980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.967995 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.968014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:48 crc kubenswrapper[4809]: I1127 17:09:48.968030 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:48Z","lastTransitionTime":"2025-11-27T17:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.070487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.070537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.070551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.070572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.070585 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.173610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.173652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.173663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.173681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.173694 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.250531 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.250684 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.250771 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:09:53.250711496 +0000 UTC m=+28.523168848 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.250844 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.250847 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.250897 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:53.25088256 +0000 UTC m=+28.523339912 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.250929 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.251044 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.251119 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.251139 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.251152 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.251165 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:53.251142587 +0000 UTC m=+28.523599939 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.251196 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:53.251186158 +0000 UTC m=+28.523643630 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.276230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.276272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.276282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.276298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.276308 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.302620 4809 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.351486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.351721 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.351775 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.351787 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.351852 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 17:09:53.351835358 +0000 UTC m=+28.624292710 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.378750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.378787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.378798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.378816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.378827 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.457069 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.457123 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.457098 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.457260 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.457403 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:09:49 crc kubenswrapper[4809]: E1127 17:09:49.457609 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.481359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.481398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.481410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.481427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.481441 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.584450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.584483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.584491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.584505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.584514 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.647526 4809 generic.go:334] "Generic (PLEG): container finished" podID="21848912-a128-4704-b84a-811af21e99cc" containerID="5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392" exitCode=0 Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.647601 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerDied","Data":"5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.649072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.655357 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.662532 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.680370 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.686551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.686583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.686595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.686613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.686627 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.692163 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.707949 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.722277 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.735778 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.748632 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.762510 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.775550 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.789128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.789170 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.789181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.789204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.789217 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.789365 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.802555 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.816508 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.831176 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.844690 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.863159 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.875589 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.891668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.891715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.891727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.891761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.891775 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.892161 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.905753 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.919721 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.929771 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.942983 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.954101 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.967780 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.979483 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.993863 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:49Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.994676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.994712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.994723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.994774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:49 crc kubenswrapper[4809]: I1127 17:09:49.994789 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:49Z","lastTransitionTime":"2025-11-27T17:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.004453 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.018504 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.030080 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.097182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.097219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.097229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.097247 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.097256 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:50Z","lastTransitionTime":"2025-11-27T17:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.199529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.199553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.199561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.199574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.199582 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:50Z","lastTransitionTime":"2025-11-27T17:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.244840 4809 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.301889 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.301916 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.301924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.301937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.301947 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:50Z","lastTransitionTime":"2025-11-27T17:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.404195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.404225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.404233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.404246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.404255 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:50Z","lastTransitionTime":"2025-11-27T17:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.507202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.507245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.507266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.507285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.507298 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:50Z","lastTransitionTime":"2025-11-27T17:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.610145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.610175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.610183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.610197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.610207 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:50Z","lastTransitionTime":"2025-11-27T17:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.660112 4809 generic.go:334] "Generic (PLEG): container finished" podID="21848912-a128-4704-b84a-811af21e99cc" containerID="d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121" exitCode=0 Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.660190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerDied","Data":"d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.677199 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.696950 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.707929 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.712124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.712167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.712208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.712226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.712895 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:50Z","lastTransitionTime":"2025-11-27T17:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.720710 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.733805 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.745811 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.758134 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.773132 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.786095 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.797915 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.811432 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.816352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.816409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.816433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.816454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.816466 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:50Z","lastTransitionTime":"2025-11-27T17:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.823902 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.835244 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.847618 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:50Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.918818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.918860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.918871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.918887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:50 crc kubenswrapper[4809]: I1127 17:09:50.918897 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:50Z","lastTransitionTime":"2025-11-27T17:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.021768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.022794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.022810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.022825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.022835 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.125368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.125424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.125440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.125461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.125476 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.227571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.227604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.227615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.227635 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.227647 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.330137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.330202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.330218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.330241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.330260 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.432806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.432850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.432859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.432876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.432887 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.457481 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.457598 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:51 crc kubenswrapper[4809]: E1127 17:09:51.457683 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.457701 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:51 crc kubenswrapper[4809]: E1127 17:09:51.457882 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:09:51 crc kubenswrapper[4809]: E1127 17:09:51.458015 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.535239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.535277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.535285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.535302 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.535311 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.638341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.638393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.638409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.638431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.638447 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.664993 4809 generic.go:334] "Generic (PLEG): container finished" podID="21848912-a128-4704-b84a-811af21e99cc" containerID="b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64" exitCode=0 Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.665033 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerDied","Data":"b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.679604 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.692638 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.708867 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.725954 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.741245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.741280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.741288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.741303 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.741313 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.744616 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.756470 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.770896 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.782395 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.796003 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.808803 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.822125 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.834164 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.843621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.843657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.843666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.843680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.843689 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.846209 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.860108 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:51Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.945442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.945475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.945485 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.945500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:51 crc kubenswrapper[4809]: I1127 17:09:51.945510 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:51Z","lastTransitionTime":"2025-11-27T17:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.047591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.047631 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.047642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.047656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.047666 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.154033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.154078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.154092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.154114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.154126 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.256982 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.257064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.257084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.257114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.257139 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.359955 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.359994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.360003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.360020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.360031 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.462965 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.463276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.463291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.463309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.463321 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.566216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.566281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.566295 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.566317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.566334 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.669104 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.669171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.669187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.669216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.669233 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.675190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.675788 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.675949 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.680653 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerStarted","Data":"80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.693972 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.710926 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.711478 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.716218 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.732818 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.750552 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.766653 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.771841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.771883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.771894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.771913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.771924 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.781086 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.795335 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.811300 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.826702 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.846900 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.861009 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.873818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.873874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.873888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.873908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.873923 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.874797 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.886984 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.900643 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.913099 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.926277 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.941783 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.956017 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.973387 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.976539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.976591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.976607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.976630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.976647 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:52Z","lastTransitionTime":"2025-11-27T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:52 crc kubenswrapper[4809]: I1127 17:09:52.991894 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:52Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.008389 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.025283 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.042672 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.057371 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.074857 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.084260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.084318 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.084329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.084350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.084361 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:53Z","lastTransitionTime":"2025-11-27T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.092492 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.110550 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.134398 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.186494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.186536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.186546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.186564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.186577 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:53Z","lastTransitionTime":"2025-11-27T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.288235 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.288411 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:10:01.288387908 +0000 UTC m=+36.560845260 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.288418 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.288573 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.288634 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:01.288623714 +0000 UTC m=+36.561081066 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.289133 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.289216 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.289249 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:01.289241061 +0000 UTC m=+36.561698413 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.289251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.289379 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.289392 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.289403 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.289468 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:01.289461148 +0000 UTC m=+36.561918500 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.289953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.290003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.290019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.290050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.290068 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:53Z","lastTransitionTime":"2025-11-27T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.391203 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.391509 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.391547 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.391568 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.391788 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:01.391731191 +0000 UTC m=+36.664188563 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.392366 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.392394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.392403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.392419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.392431 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:53Z","lastTransitionTime":"2025-11-27T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.457277 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.457308 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.457460 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.457308 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.457550 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.457756 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.494858 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.494933 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.494972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.494996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.495008 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:53Z","lastTransitionTime":"2025-11-27T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.597431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.597477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.597487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.597510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.597523 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:53Z","lastTransitionTime":"2025-11-27T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.687412 4809 generic.go:334] "Generic (PLEG): container finished" podID="21848912-a128-4704-b84a-811af21e99cc" containerID="80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e" exitCode=0 Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.687577 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.688389 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerDied","Data":"80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.699671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.699715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.699727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.699770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.699785 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:53Z","lastTransitionTime":"2025-11-27T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.704579 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.718435 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.734600 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.748326 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.790509 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.802728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.802773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.802782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.802795 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.802806 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:53Z","lastTransitionTime":"2025-11-27T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.819086 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.834098 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.846642 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.863239 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.883640 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.895580 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.905350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.905399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.905412 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.905430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.905442 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:53Z","lastTransitionTime":"2025-11-27T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.909583 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.921484 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.936108 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:53Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.970240 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:09:53 crc kubenswrapper[4809]: I1127 17:09:53.970927 4809 scope.go:117] "RemoveContainer" containerID="13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847" Nov 27 17:09:53 crc kubenswrapper[4809]: E1127 17:09:53.971071 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.008107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.008159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.008175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.008193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.008205 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.114984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.115054 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.115065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.115081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.115092 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.217871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.217912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.217921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.217937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.217946 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.321146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.321199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.321217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.321241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.321253 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.424455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.424507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.424520 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.424538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.424550 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.527069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.527125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.527139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.527158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.527170 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.630229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.630283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.630296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.630315 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.630329 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.694997 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.696175 4809 generic.go:334] "Generic (PLEG): container finished" podID="21848912-a128-4704-b84a-811af21e99cc" containerID="ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53" exitCode=0 Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.696822 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerDied","Data":"ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.715006 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.732219 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.733615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.733654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.733662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.733682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.733694 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.754158 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.774929 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.796853 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.814646 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.830313 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.837162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.837225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.837248 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.837278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.837298 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.850310 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.869716 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.885459 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.904036 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.929408 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.939773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.939945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.940074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.940207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.940290 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:54Z","lastTransitionTime":"2025-11-27T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.941991 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:54 crc kubenswrapper[4809]: I1127 17:09:54.966093 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:54Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.043764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.043820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.043834 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.043855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.043866 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.046258 4809 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.116941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.116999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.117013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.117034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.117052 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: E1127 17:09:55.131491 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.136008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.136055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.136069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.136093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.136106 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: E1127 17:09:55.153505 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.157847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.157886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.157895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.157912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.157925 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: E1127 17:09:55.172424 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.177715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.177800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.177810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.177826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.177839 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: E1127 17:09:55.190794 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.195291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.195712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.195759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.195783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.195798 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: E1127 17:09:55.212432 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: E1127 17:09:55.212562 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.214381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.214418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.214429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.214444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.214454 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.317224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.317262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.317271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.317286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.317295 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.419484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.419528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.419546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.419564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.419576 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.457141 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.457295 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.457444 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:55 crc kubenswrapper[4809]: E1127 17:09:55.457429 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:09:55 crc kubenswrapper[4809]: E1127 17:09:55.457532 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:09:55 crc kubenswrapper[4809]: E1127 17:09:55.457613 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.474632 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.490315 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.505983 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.519266 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.522252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.522275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.522284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.522299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.522309 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.534037 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.550104 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.561956 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.578331 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.601719 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.625021 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.625071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.625081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.625099 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.625109 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.625509 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.639857 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.655788 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.668377 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.682307 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.710324 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" event={"ID":"21848912-a128-4704-b84a-811af21e99cc","Type":"ContainerStarted","Data":"31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb"} Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.723700 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.728311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.728339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.728348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.728364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.728374 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.736960 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.749726 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.767215 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.777622 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.793201 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.807905 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.823649 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.830881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.830929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.830961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.830987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.831001 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.836169 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.849557 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.864013 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.883569 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.898754 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.915683 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.934438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.934518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.934536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.934565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:55 crc kubenswrapper[4809]: I1127 17:09:55.934582 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:55Z","lastTransitionTime":"2025-11-27T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.037138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.037188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.037200 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.037221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.037232 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.140051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.140119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.140131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.140152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.140167 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.243861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.243913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.243926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.243947 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.243963 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.347008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.347055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.347065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.347081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.347094 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.450080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.450467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.450533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.450661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.450799 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.461026 4809 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.553855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.553931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.553949 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.553978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.553998 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.657277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.657327 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.657379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.657408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.657423 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.721636 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/0.log" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.725362 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae" exitCode=1 Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.725407 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.726148 4809 scope.go:117] "RemoveContainer" containerID="af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.742339 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.755761 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.759589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.759627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.759638 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.759656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.759669 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.770057 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.786411 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.807701 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"roller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688119 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688291 6038 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688503 6038 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.688700 6038 factory.go:656] Stopping watch factory\\\\nI1127 17:09:55.689109 6038 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.689154 6038 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.689276 6038 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1127 17:09:55.689367 6038 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.818411 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.831723 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.844834 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.856448 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.862306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.862338 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.862348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.862365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.862376 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.870561 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.884622 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.898299 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.912369 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.923928 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.965349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.965392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.965401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.965420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:56 crc kubenswrapper[4809]: I1127 17:09:56.965431 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:56Z","lastTransitionTime":"2025-11-27T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.068757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.068813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.068824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.068842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.068861 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.171994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.172471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.172483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.172503 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.172519 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.274328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.274365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.274374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.274388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.274399 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.378114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.378180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.378191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.378214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.378226 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.457125 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.457178 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.457233 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:57 crc kubenswrapper[4809]: E1127 17:09:57.457289 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:09:57 crc kubenswrapper[4809]: E1127 17:09:57.457424 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:09:57 crc kubenswrapper[4809]: E1127 17:09:57.457548 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.480758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.480810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.480822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.480839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.480849 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.583525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.583583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.583592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.583615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.583628 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.686449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.686490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.686502 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.686520 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.686533 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.730170 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/0.log" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.733433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.734026 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.749723 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.763048 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.775491 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.789042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.789097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.789115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.789140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.789161 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.789484 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.802882 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.816880 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.827913 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.842182 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.859419 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.877934 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"roller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688119 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688291 6038 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688503 6038 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.688700 6038 factory.go:656] Stopping watch factory\\\\nI1127 17:09:55.689109 6038 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.689154 6038 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.689276 6038 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1127 17:09:55.689367 6038 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.891195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.891243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.891253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.891273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.891288 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.894591 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.906221 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.918349 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c"] Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.918367 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.918999 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.920999 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.921472 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.933962 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.945402 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.959611 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.969466 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.981812 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.993408 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.993580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.993616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.993632 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.993651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:57 crc kubenswrapper[4809]: I1127 17:09:57.993664 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:57Z","lastTransitionTime":"2025-11-27T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.005712 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.017720 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.030958 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.038523 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwzt5\" (UniqueName: \"kubernetes.io/projected/afa276bf-d508-44a3-bde0-c9a96a4de65d-kube-api-access-dwzt5\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.038562 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/afa276bf-d508-44a3-bde0-c9a96a4de65d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.038591 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/afa276bf-d508-44a3-bde0-c9a96a4de65d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.038649 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/afa276bf-d508-44a3-bde0-c9a96a4de65d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.040052 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.051358 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.060626 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.071715 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.090149 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"roller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688119 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688291 6038 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688503 6038 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.688700 6038 factory.go:656] Stopping watch factory\\\\nI1127 17:09:55.689109 6038 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.689154 6038 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.689276 6038 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1127 17:09:55.689367 6038 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.096004 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.096029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.096038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.096053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.096064 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:58Z","lastTransitionTime":"2025-11-27T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.107851 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.121360 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.140280 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwzt5\" (UniqueName: \"kubernetes.io/projected/afa276bf-d508-44a3-bde0-c9a96a4de65d-kube-api-access-dwzt5\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.140378 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/afa276bf-d508-44a3-bde0-c9a96a4de65d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.140441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/afa276bf-d508-44a3-bde0-c9a96a4de65d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.140491 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/afa276bf-d508-44a3-bde0-c9a96a4de65d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.141035 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/afa276bf-d508-44a3-bde0-c9a96a4de65d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.142225 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/afa276bf-d508-44a3-bde0-c9a96a4de65d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.151591 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/afa276bf-d508-44a3-bde0-c9a96a4de65d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.158656 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwzt5\" (UniqueName: \"kubernetes.io/projected/afa276bf-d508-44a3-bde0-c9a96a4de65d-kube-api-access-dwzt5\") pod \"ovnkube-control-plane-749d76644c-pt22c\" (UID: \"afa276bf-d508-44a3-bde0-c9a96a4de65d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.198459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.198507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.198523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.198543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.198557 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:58Z","lastTransitionTime":"2025-11-27T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.234909 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" Nov 27 17:09:58 crc kubenswrapper[4809]: W1127 17:09:58.246626 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafa276bf_d508_44a3_bde0_c9a96a4de65d.slice/crio-d915684807ab10d182e07240432b8aa2fdc55e064f1ca43c6d3ebc32bc12b7a8 WatchSource:0}: Error finding container d915684807ab10d182e07240432b8aa2fdc55e064f1ca43c6d3ebc32bc12b7a8: Status 404 returned error can't find the container with id d915684807ab10d182e07240432b8aa2fdc55e064f1ca43c6d3ebc32bc12b7a8 Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.300608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.300648 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.300657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.300672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.300682 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:58Z","lastTransitionTime":"2025-11-27T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.402997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.403036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.403045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.403070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.403082 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:58Z","lastTransitionTime":"2025-11-27T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.505601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.505661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.505679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.505703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.505719 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:58Z","lastTransitionTime":"2025-11-27T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.608399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.608441 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.608453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.608472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.608484 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:58Z","lastTransitionTime":"2025-11-27T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.711598 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.711658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.711675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.711699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.711715 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:58Z","lastTransitionTime":"2025-11-27T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.738376 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/1.log" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.739164 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/0.log" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.742456 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656" exitCode=1 Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.742516 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.742558 4809 scope.go:117] "RemoveContainer" containerID="af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.743619 4809 scope.go:117] "RemoveContainer" containerID="d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656" Nov 27 17:09:58 crc kubenswrapper[4809]: E1127 17:09:58.743913 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.744677 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" event={"ID":"afa276bf-d508-44a3-bde0-c9a96a4de65d","Type":"ContainerStarted","Data":"4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.744729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" event={"ID":"afa276bf-d508-44a3-bde0-c9a96a4de65d","Type":"ContainerStarted","Data":"2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.744748 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" event={"ID":"afa276bf-d508-44a3-bde0-c9a96a4de65d","Type":"ContainerStarted","Data":"d915684807ab10d182e07240432b8aa2fdc55e064f1ca43c6d3ebc32bc12b7a8"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.763145 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.791142 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"roller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688119 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688291 6038 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688503 6038 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.688700 6038 factory.go:656] Stopping watch factory\\\\nI1127 17:09:55.689109 6038 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.689154 6038 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.689276 6038 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1127 17:09:55.689367 6038 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"y.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1127 17:09:57.548903 6241 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 17:09:57.548904 6241 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 17:09:57.548908 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1127 17:09:57.548973 6241 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:09:57.548624 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.804949 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.813878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.813928 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.813940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.813960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.813972 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:58Z","lastTransitionTime":"2025-11-27T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.818982 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.842089 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.859418 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.874008 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.886398 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.899896 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.913296 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.916612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.916656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.916672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.916695 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.916710 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:58Z","lastTransitionTime":"2025-11-27T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.927681 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.941900 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.955411 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.970262 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.983127 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.998842 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-2nc2d"] Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.999441 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:09:58 crc kubenswrapper[4809]: E1127 17:09:58.999532 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:09:58 crc kubenswrapper[4809]: I1127 17:09:58.999787 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:58Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.012158 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.019072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.019128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.019141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.019163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.019178 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.031378 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"roller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688119 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688291 6038 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688503 6038 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.688700 6038 factory.go:656] Stopping watch factory\\\\nI1127 17:09:55.689109 6038 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.689154 6038 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.689276 6038 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1127 17:09:55.689367 6038 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"y.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1127 17:09:57.548903 6241 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 17:09:57.548904 6241 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 17:09:57.548908 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1127 17:09:57.548973 6241 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:09:57.548624 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.047779 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.048112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79cvs\" (UniqueName: \"kubernetes.io/projected/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-kube-api-access-79cvs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.048235 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.060956 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.080497 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.117027 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.122009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.122061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.122073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.122094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.122112 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.140728 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.149596 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.149680 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79cvs\" (UniqueName: \"kubernetes.io/projected/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-kube-api-access-79cvs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:09:59 crc kubenswrapper[4809]: E1127 17:09:59.149892 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:09:59 crc kubenswrapper[4809]: E1127 17:09:59.150000 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs podName:683f4d0e-67aa-40a7-b7b0-e5325ed0225f nodeName:}" failed. No retries permitted until 2025-11-27 17:09:59.649972704 +0000 UTC m=+34.922430066 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs") pod "network-metrics-daemon-2nc2d" (UID: "683f4d0e-67aa-40a7-b7b0-e5325ed0225f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.159983 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.170924 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79cvs\" (UniqueName: \"kubernetes.io/projected/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-kube-api-access-79cvs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.172481 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.183399 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.194270 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.205601 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.214442 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.223904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.223953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.223965 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.223985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.223998 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.225044 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.237970 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.251354 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.262395 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.273374 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.284252 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.297717 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.310944 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.326622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.326668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.326678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.326694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.326706 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.336691 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.351724 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.364542 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.381209 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.402925 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af18fdd9af52118b3194076807d383f0e13a1c91d65075bd35c9c89e501cbeae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"message\\\":\\\"roller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688119 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688291 6038 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.688503 6038 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.688700 6038 factory.go:656] Stopping watch factory\\\\nI1127 17:09:55.689109 6038 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 17:09:55.689154 6038 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 17:09:55.689276 6038 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1127 17:09:55.689367 6038 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"y.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1127 17:09:57.548903 6241 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 17:09:57.548904 6241 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 17:09:57.548908 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1127 17:09:57.548973 6241 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:09:57.548624 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.417101 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.429931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.429984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.430000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.430020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.430033 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.432082 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.443965 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.455775 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.457066 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:09:59 crc kubenswrapper[4809]: E1127 17:09:59.457193 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.457270 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:09:59 crc kubenswrapper[4809]: E1127 17:09:59.457478 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.457618 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:09:59 crc kubenswrapper[4809]: E1127 17:09:59.457741 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.532580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.532618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.532627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.532642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.532652 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.635402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.635448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.635457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.635473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.635485 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.655102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:09:59 crc kubenswrapper[4809]: E1127 17:09:59.655284 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:09:59 crc kubenswrapper[4809]: E1127 17:09:59.655358 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs podName:683f4d0e-67aa-40a7-b7b0-e5325ed0225f nodeName:}" failed. No retries permitted until 2025-11-27 17:10:00.655339728 +0000 UTC m=+35.927797080 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs") pod "network-metrics-daemon-2nc2d" (UID: "683f4d0e-67aa-40a7-b7b0-e5325ed0225f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.738633 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.738689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.738697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.738713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.738723 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.748369 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/1.log" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.752467 4809 scope.go:117] "RemoveContainer" containerID="d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656" Nov 27 17:09:59 crc kubenswrapper[4809]: E1127 17:09:59.752631 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.765066 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.777713 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.801278 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"y.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1127 17:09:57.548903 6241 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 17:09:57.548904 6241 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 17:09:57.548908 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1127 17:09:57.548973 6241 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:09:57.548624 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.813722 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.823232 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.834088 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.841159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.841221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.841237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.841255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.841288 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.846150 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.860125 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.875816 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.891299 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.904898 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.921030 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.934409 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.947477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.947818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.948138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.948164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.948179 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:09:59Z","lastTransitionTime":"2025-11-27T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.949728 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.965088 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:09:59 crc kubenswrapper[4809]: I1127 17:09:59.981152 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:59Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.051369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.051422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.051435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.051455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.051470 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.154835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.154894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.154904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.154923 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.154935 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.257346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.257393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.257407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.257424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.257440 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.359796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.359841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.359878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.359899 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.359911 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.457831 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:00 crc kubenswrapper[4809]: E1127 17:10:00.458023 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.462445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.462482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.462494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.462512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.462524 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.563977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.564020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.564033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.564050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.564063 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.664046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:00 crc kubenswrapper[4809]: E1127 17:10:00.664190 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:00 crc kubenswrapper[4809]: E1127 17:10:00.664245 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs podName:683f4d0e-67aa-40a7-b7b0-e5325ed0225f nodeName:}" failed. No retries permitted until 2025-11-27 17:10:02.664230176 +0000 UTC m=+37.936687528 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs") pod "network-metrics-daemon-2nc2d" (UID: "683f4d0e-67aa-40a7-b7b0-e5325ed0225f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.665827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.665861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.665873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.665891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.665903 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.768178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.768225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.768238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.768256 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.768271 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.871280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.871686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.871698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.871715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.871725 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.975675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.975768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.975781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.975805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:00 crc kubenswrapper[4809]: I1127 17:10:00.975818 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:00Z","lastTransitionTime":"2025-11-27T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.079431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.079511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.079534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.079563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.079585 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:01Z","lastTransitionTime":"2025-11-27T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.183496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.183543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.183553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.183572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.183584 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:01Z","lastTransitionTime":"2025-11-27T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.287634 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.287715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.287746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.287841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.287866 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:01Z","lastTransitionTime":"2025-11-27T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.373463 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.373672 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:10:17.373628894 +0000 UTC m=+52.646086296 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.373796 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.373972 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.374004 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.374056 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.374095 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:17.374070836 +0000 UTC m=+52.646528228 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.374244 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.374282 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.374330 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.374356 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.374413 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:17.374365845 +0000 UTC m=+52.646823247 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.374458 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:17.374437977 +0000 UTC m=+52.646895389 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.391113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.391179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.391198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.391232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.391252 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:01Z","lastTransitionTime":"2025-11-27T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.457781 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.457874 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.457784 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.457985 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.458085 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.458138 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.475906 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.476185 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.476231 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.476255 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:01 crc kubenswrapper[4809]: E1127 17:10:01.476406 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:17.476378581 +0000 UTC m=+52.748835963 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.494940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.494989 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.495000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.495018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.495030 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:01Z","lastTransitionTime":"2025-11-27T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.599639 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.599709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.599771 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.599805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.599833 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:01Z","lastTransitionTime":"2025-11-27T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.702812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.702862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.702872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.702891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.702903 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:01Z","lastTransitionTime":"2025-11-27T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.806466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.806538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.806554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.806580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.806599 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:01Z","lastTransitionTime":"2025-11-27T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.910319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.910390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.910408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.910434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:01 crc kubenswrapper[4809]: I1127 17:10:01.910454 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:01Z","lastTransitionTime":"2025-11-27T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.013008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.013068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.013084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.013105 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.013117 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.116501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.116582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.116600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.116630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.116647 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.219865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.219941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.219960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.219988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.220008 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.322565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.322634 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.322661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.322698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.322720 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.426030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.426092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.426106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.426129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.426144 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.457901 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:02 crc kubenswrapper[4809]: E1127 17:10:02.458134 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.529702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.529798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.529816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.529841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.529860 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.633444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.633494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.633542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.633573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.633587 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.689199 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:02 crc kubenswrapper[4809]: E1127 17:10:02.689402 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:02 crc kubenswrapper[4809]: E1127 17:10:02.689462 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs podName:683f4d0e-67aa-40a7-b7b0-e5325ed0225f nodeName:}" failed. No retries permitted until 2025-11-27 17:10:06.689443717 +0000 UTC m=+41.961901069 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs") pod "network-metrics-daemon-2nc2d" (UID: "683f4d0e-67aa-40a7-b7b0-e5325ed0225f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.738664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.738721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.738733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.738775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.738790 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.842341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.842399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.842408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.842422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.842432 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.945181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.945270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.945285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.945304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:02 crc kubenswrapper[4809]: I1127 17:10:02.945320 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:02Z","lastTransitionTime":"2025-11-27T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.048207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.048277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.048300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.048355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.048371 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.151133 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.151195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.151208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.151228 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.151243 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.253927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.254002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.254012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.254028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.254038 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.356472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.356535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.356553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.356580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.356599 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.457354 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.457419 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.457370 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:03 crc kubenswrapper[4809]: E1127 17:10:03.457507 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:03 crc kubenswrapper[4809]: E1127 17:10:03.457576 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:03 crc kubenswrapper[4809]: E1127 17:10:03.457946 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.459019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.459050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.459060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.459075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.459085 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.561420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.561484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.561497 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.561515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.561531 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.663777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.663818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.663830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.663848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.663859 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.765919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.765952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.765961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.765976 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.765985 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.868726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.868814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.868827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.868850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.868863 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.972024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.972087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.972101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.972126 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:03 crc kubenswrapper[4809]: I1127 17:10:03.972140 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:03Z","lastTransitionTime":"2025-11-27T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.075447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.075516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.075529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.075546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.075558 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:04Z","lastTransitionTime":"2025-11-27T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.179395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.179465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.179480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.179510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.179528 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:04Z","lastTransitionTime":"2025-11-27T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.282246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.282341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.282368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.282404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.282430 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:04Z","lastTransitionTime":"2025-11-27T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.385774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.385860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.385888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.385920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.385936 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:04Z","lastTransitionTime":"2025-11-27T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.457280 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:04 crc kubenswrapper[4809]: E1127 17:10:04.457560 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.489487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.489548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.489562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.489590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.489608 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:04Z","lastTransitionTime":"2025-11-27T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.592335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.592670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.592850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.593043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.593171 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:04Z","lastTransitionTime":"2025-11-27T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.696286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.696343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.696353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.696371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.696382 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:04Z","lastTransitionTime":"2025-11-27T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.800067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.800133 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.800150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.800177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.800203 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:04Z","lastTransitionTime":"2025-11-27T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.904300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.904448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.904475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.904507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:04 crc kubenswrapper[4809]: I1127 17:10:04.904531 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:04Z","lastTransitionTime":"2025-11-27T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.008210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.008283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.008296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.008318 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.008331 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.111717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.111823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.111842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.111872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.111891 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.215177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.215254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.215274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.215303 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.215332 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.319063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.319132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.319152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.319182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.319202 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.422391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.422471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.422496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.422528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.422552 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.457084 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.457178 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.457121 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:05 crc kubenswrapper[4809]: E1127 17:10:05.457353 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:05 crc kubenswrapper[4809]: E1127 17:10:05.457528 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:05 crc kubenswrapper[4809]: E1127 17:10:05.457712 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.476209 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.494709 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.511868 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.525132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.525186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.525197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.525217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.525228 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.529848 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.543275 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.564469 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.577520 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.592087 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.593566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.593599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.593614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.593635 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.593649 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.605845 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: E1127 17:10:05.606276 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.610163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.610204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.610217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.610257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.610268 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.623023 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: E1127 17:10:05.623169 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.629103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.629159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.629176 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.629195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.629207 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.644303 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: E1127 17:10:05.644891 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.649604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.649645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.649656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.649674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.649684 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.662077 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: E1127 17:10:05.663350 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.667653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.667689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.667704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.667727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.667766 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: E1127 17:10:05.680407 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: E1127 17:10:05.680577 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.682429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.682501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.682517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.682535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.682548 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.688611 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"y.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1127 17:09:57.548903 6241 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 17:09:57.548904 6241 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 17:09:57.548908 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1127 17:09:57.548973 6241 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:09:57.548624 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.702449 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.713448 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.725430 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:05Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.784808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.784866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.784878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.784901 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.784917 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.888512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.888589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.888611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.888643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.888665 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.992642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.992732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.993164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.993636 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:05 crc kubenswrapper[4809]: I1127 17:10:05.993711 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:05Z","lastTransitionTime":"2025-11-27T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.096342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.096381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.096393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.096410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.096420 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:06Z","lastTransitionTime":"2025-11-27T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.199823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.199881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.199892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.199912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.199923 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:06Z","lastTransitionTime":"2025-11-27T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.302637 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.302688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.302699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.302718 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.302731 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:06Z","lastTransitionTime":"2025-11-27T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.405453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.405487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.405496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.405510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.405519 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:06Z","lastTransitionTime":"2025-11-27T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.457572 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:06 crc kubenswrapper[4809]: E1127 17:10:06.457925 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.508641 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.508693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.508706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.508730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.508760 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:06Z","lastTransitionTime":"2025-11-27T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.611897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.611945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.611957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.611977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.611989 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:06Z","lastTransitionTime":"2025-11-27T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.715214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.715272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.715290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.715317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.715336 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:06Z","lastTransitionTime":"2025-11-27T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.737517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:06 crc kubenswrapper[4809]: E1127 17:10:06.737827 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:06 crc kubenswrapper[4809]: E1127 17:10:06.737982 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs podName:683f4d0e-67aa-40a7-b7b0-e5325ed0225f nodeName:}" failed. No retries permitted until 2025-11-27 17:10:14.737948875 +0000 UTC m=+50.010406267 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs") pod "network-metrics-daemon-2nc2d" (UID: "683f4d0e-67aa-40a7-b7b0-e5325ed0225f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.818419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.818508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.818528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.818559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.818583 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:06Z","lastTransitionTime":"2025-11-27T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.922193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.922270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.922296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.922330 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:06 crc kubenswrapper[4809]: I1127 17:10:06.922355 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:06Z","lastTransitionTime":"2025-11-27T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.025825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.025871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.025882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.025908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.025922 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.128680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.128760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.128778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.128799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.128811 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.231337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.231420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.231435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.231459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.231479 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.334547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.334614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.334626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.334644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.334655 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.437950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.438015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.438038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.438064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.438083 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.457947 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.458069 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.458135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:07 crc kubenswrapper[4809]: E1127 17:10:07.458226 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:07 crc kubenswrapper[4809]: E1127 17:10:07.458787 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:07 crc kubenswrapper[4809]: E1127 17:10:07.458867 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.541879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.541924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.541933 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.541949 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.541958 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.645575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.645623 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.645634 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.645658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.645669 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.748474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.748519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.748528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.748545 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.748556 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.851375 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.851415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.851426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.851442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.851453 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.954494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.954553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.954567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.954592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:07 crc kubenswrapper[4809]: I1127 17:10:07.954604 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:07Z","lastTransitionTime":"2025-11-27T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.058363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.058427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.058439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.058461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.058474 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.161578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.161653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.161670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.161694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.161707 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.264850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.264936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.264956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.264988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.265008 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.368036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.368080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.368089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.368107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.368120 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.457756 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:08 crc kubenswrapper[4809]: E1127 17:10:08.457999 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.458838 4809 scope.go:117] "RemoveContainer" containerID="13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.470953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.470987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.470996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.471015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.471026 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.573686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.573729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.573766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.573783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.573794 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.677073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.677131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.677145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.677167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.677184 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.780237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.780293 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.780306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.780326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.780341 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.790204 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.792777 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.793615 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.811356 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.828456 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.846160 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.862626 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.883833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.883885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.883900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.883926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.883942 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.885426 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.898251 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.914309 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.934369 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.949678 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.967490 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.988240 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.988298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.988312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.988336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.988353 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:08Z","lastTransitionTime":"2025-11-27T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:08 crc kubenswrapper[4809]: I1127 17:10:08.998298 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"y.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1127 17:09:57.548903 6241 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 17:09:57.548904 6241 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 17:09:57.548908 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1127 17:09:57.548973 6241 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:09:57.548624 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:08Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.013629 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:09Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.034035 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:09Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.048387 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:09Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.061638 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:09Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.075517 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:09Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.090912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.090958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.090967 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.090985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.091000 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:09Z","lastTransitionTime":"2025-11-27T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.194325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.194394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.194411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.194437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.194456 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:09Z","lastTransitionTime":"2025-11-27T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.296865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.296917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.296926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.296945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.296955 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:09Z","lastTransitionTime":"2025-11-27T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.400650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.400703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.400715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.400762 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.400775 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:09Z","lastTransitionTime":"2025-11-27T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.457575 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.457701 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:09 crc kubenswrapper[4809]: E1127 17:10:09.457780 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.457589 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:09 crc kubenswrapper[4809]: E1127 17:10:09.457971 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:09 crc kubenswrapper[4809]: E1127 17:10:09.458159 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.504165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.504229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.504245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.504268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.504291 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:09Z","lastTransitionTime":"2025-11-27T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.608096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.608151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.608165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.608186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.608202 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:09Z","lastTransitionTime":"2025-11-27T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.712716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.712784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.712796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.712817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.712829 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:09Z","lastTransitionTime":"2025-11-27T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.816037 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.816084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.816095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.816112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.816124 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:09Z","lastTransitionTime":"2025-11-27T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.918998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.919096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.919120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.919150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:09 crc kubenswrapper[4809]: I1127 17:10:09.919171 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:09Z","lastTransitionTime":"2025-11-27T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.022560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.022636 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.022658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.022687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.022707 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.126177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.126276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.126328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.126359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.126380 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.229864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.229919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.229929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.229946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.229958 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.333434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.333486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.333505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.333528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.333547 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.437122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.437225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.437249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.437285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.437309 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.457841 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:10 crc kubenswrapper[4809]: E1127 17:10:10.458090 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.540527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.540611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.540649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.540686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.540715 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.644135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.644184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.644198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.644220 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.644235 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.747008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.747082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.747101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.747130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.747149 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.850679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.850787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.850806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.850838 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.850862 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.954367 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.954443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.954457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.954478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:10 crc kubenswrapper[4809]: I1127 17:10:10.954493 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:10Z","lastTransitionTime":"2025-11-27T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.057335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.057406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.057420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.057440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.057454 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.160695 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.160790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.160806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.160829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.160843 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.264408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.264488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.264516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.264548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.264568 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.367556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.367630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.367649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.367674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.367692 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.457280 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.457314 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.457557 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:11 crc kubenswrapper[4809]: E1127 17:10:11.457512 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:11 crc kubenswrapper[4809]: E1127 17:10:11.458085 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:11 crc kubenswrapper[4809]: E1127 17:10:11.458186 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.458485 4809 scope.go:117] "RemoveContainer" containerID="d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.470580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.470655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.470680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.470710 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.470733 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.573999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.574082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.574107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.574140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.574166 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.677608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.677992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.678011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.678039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.678058 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.781661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.781712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.781724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.781770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.781785 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.807174 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/1.log" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.809088 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.809512 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.826773 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.845864 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"y.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1127 17:09:57.548903 6241 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 17:09:57.548904 6241 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 17:09:57.548908 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1127 17:09:57.548973 6241 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:09:57.548624 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.863709 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.880017 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.884149 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.884188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.884203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.884221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.884232 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.893687 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.911149 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.936283 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.953772 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.968001 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.979910 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.989586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.989639 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.989664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.989688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.989701 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:11Z","lastTransitionTime":"2025-11-27T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:11 crc kubenswrapper[4809]: I1127 17:10:11.991757 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:11Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.005125 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.021501 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.044298 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.065260 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.081511 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.091760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.091797 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.091810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.091826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.091838 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:12Z","lastTransitionTime":"2025-11-27T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.195018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.195066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.195079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.195100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.195115 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:12Z","lastTransitionTime":"2025-11-27T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.298434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.298491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.298507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.298527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.298569 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:12Z","lastTransitionTime":"2025-11-27T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.402363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.402430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.402448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.402476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.402496 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:12Z","lastTransitionTime":"2025-11-27T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.457512 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:12 crc kubenswrapper[4809]: E1127 17:10:12.457760 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.506669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.506731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.506810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.506866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.506972 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:12Z","lastTransitionTime":"2025-11-27T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.610685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.610810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.610839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.610872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.610900 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:12Z","lastTransitionTime":"2025-11-27T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.714690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.714783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.714800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.714823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.714834 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:12Z","lastTransitionTime":"2025-11-27T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.815982 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/2.log" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.816556 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/1.log" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.817405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.817441 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.817449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.817465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.817477 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:12Z","lastTransitionTime":"2025-11-27T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.820461 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c" exitCode=1 Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.820520 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.820570 4809 scope.go:117] "RemoveContainer" containerID="d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.821229 4809 scope.go:117] "RemoveContainer" containerID="51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c" Nov 27 17:10:12 crc kubenswrapper[4809]: E1127 17:10:12.821414 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.843040 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.869917 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6fc815f72cd4c475f84a77362c739d1e89493b9063d8ecc21391ef2427c9656\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"y.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1127 17:09:57.548903 6241 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 17:09:57.548904 6241 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 17:09:57.548908 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1127 17:09:57.548973 6241 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:09:57Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:09:57.548624 6241 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/net\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.887809 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.906052 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.919824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.919863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.919873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.919892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.919916 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:12Z","lastTransitionTime":"2025-11-27T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.921962 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.935729 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.949221 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.966568 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.981595 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:12 crc kubenswrapper[4809]: I1127 17:10:12.996863 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.016501 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.023538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.023600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.023614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.023637 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.023656 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.036875 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.054473 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.078477 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.095699 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.112184 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.127523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.127579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.127591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.127610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.127623 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.230222 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.230658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.230767 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.230874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.230958 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.334341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.334659 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.334784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.334903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.334973 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.437613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.437664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.437676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.437697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.437708 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.457480 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.457511 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.457491 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:13 crc kubenswrapper[4809]: E1127 17:10:13.457609 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:13 crc kubenswrapper[4809]: E1127 17:10:13.457689 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:13 crc kubenswrapper[4809]: E1127 17:10:13.457769 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.539932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.539981 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.539994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.540010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.540020 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.643425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.643483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.643498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.643520 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.643535 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.746840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.746929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.746953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.746979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.746997 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.825273 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/2.log" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.829239 4809 scope.go:117] "RemoveContainer" containerID="51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c" Nov 27 17:10:13 crc kubenswrapper[4809]: E1127 17:10:13.829441 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.847230 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.849354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.849400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.849417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.849439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.849456 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.861370 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.882787 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.903153 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.920107 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.942940 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.953127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.953182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.953204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.953232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.953252 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:13Z","lastTransitionTime":"2025-11-27T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.967450 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:13 crc kubenswrapper[4809]: I1127 17:10:13.987716 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:13Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.009090 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:14Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.025337 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:14Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.040160 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:14Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.055144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.055179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.055189 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.055204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.055215 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.057028 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:14Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.081756 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:14Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.098708 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:14Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.115093 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:14Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.130807 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:14Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.158119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.158187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.158211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.158240 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.158259 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.261391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.261450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.261463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.261486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.261499 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.364610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.364686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.364708 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.364792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.364832 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.457945 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:14 crc kubenswrapper[4809]: E1127 17:10:14.458175 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.467926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.468103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.468201 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.468291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.468359 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.571631 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.571703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.571721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.571787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.571807 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.675239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.675307 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.675324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.675352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.675369 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.740283 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:14 crc kubenswrapper[4809]: E1127 17:10:14.740615 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:14 crc kubenswrapper[4809]: E1127 17:10:14.740767 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs podName:683f4d0e-67aa-40a7-b7b0-e5325ed0225f nodeName:}" failed. No retries permitted until 2025-11-27 17:10:30.740701619 +0000 UTC m=+66.013159001 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs") pod "network-metrics-daemon-2nc2d" (UID: "683f4d0e-67aa-40a7-b7b0-e5325ed0225f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.780049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.780143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.780188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.780224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.780246 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.884375 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.884425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.884435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.884453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.884463 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.987134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.987173 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.987182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.987197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:14 crc kubenswrapper[4809]: I1127 17:10:14.987208 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:14Z","lastTransitionTime":"2025-11-27T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.090402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.090456 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.090467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.090487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.090498 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.193175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.193215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.193225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.193241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.193252 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.296784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.296874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.296888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.296910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.296927 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.400430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.400492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.400503 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.400522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.400535 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.457438 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.457568 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:15 crc kubenswrapper[4809]: E1127 17:10:15.457604 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:15 crc kubenswrapper[4809]: E1127 17:10:15.457816 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.457956 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:15 crc kubenswrapper[4809]: E1127 17:10:15.458203 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.477477 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.503494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.503570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.503586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.503608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.503620 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.509475 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.530288 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.550854 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.566345 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.586814 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.602122 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.606021 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.606057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.606068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.606085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.606096 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.618399 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.634712 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.652841 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.668939 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.687902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.687966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.687979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.688000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.688016 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.688483 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.702588 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: E1127 17:10:15.703590 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.711516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.711576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.711595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.711621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.711639 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.726329 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: E1127 17:10:15.735350 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.741182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.741331 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.741402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.741481 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.741538 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.743231 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.757233 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: E1127 17:10:15.759684 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.764560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.764621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.764636 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.764663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.764685 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: E1127 17:10:15.779561 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.783891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.783934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.783945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.783964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.783976 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: E1127 17:10:15.798710 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:15Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:15 crc kubenswrapper[4809]: E1127 17:10:15.798837 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.800323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.800389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.800408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.800432 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.800450 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.903040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.903091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.903103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.903123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:15 crc kubenswrapper[4809]: I1127 17:10:15.903135 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:15Z","lastTransitionTime":"2025-11-27T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.005972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.006006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.006016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.006031 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.006041 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.108975 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.109051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.109071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.109102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.109120 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.212678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.212778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.212804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.212832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.212850 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.317075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.317158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.317184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.317216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.317240 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.420374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.420462 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.420494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.420525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.420551 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.456941 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:16 crc kubenswrapper[4809]: E1127 17:10:16.457173 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.523081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.523125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.523136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.523157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.523170 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.626886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.626932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.626940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.626957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.626971 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.730654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.730722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.730788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.730828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.730853 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.833623 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.833711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.833773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.833812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.833837 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.936626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.936683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.936697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.936717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:16 crc kubenswrapper[4809]: I1127 17:10:16.936728 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:16Z","lastTransitionTime":"2025-11-27T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.040494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.040565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.040577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.040601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.040614 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.143439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.143493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.143505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.143531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.143544 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.246487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.246970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.247107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.247271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.247415 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.351112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.351592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.351848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.352139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.352325 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.455294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.455361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.455380 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.455406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.455422 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.457867 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.457958 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.458097 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.458111 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.458306 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.458539 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.474491 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.474778 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.474817 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:10:49.474769993 +0000 UTC m=+84.747227385 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.474896 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.474955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.475009 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.475079 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.475131 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:49.475103042 +0000 UTC m=+84.747560394 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.475133 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.475162 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:49.475146383 +0000 UTC m=+84.747603765 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.475176 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.475198 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.475271 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:49.475251936 +0000 UTC m=+84.747709328 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.558599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.558657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.558667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.558687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.558698 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.576753 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.577122 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.577169 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.577184 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:17 crc kubenswrapper[4809]: E1127 17:10:17.577279 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 17:10:49.577250542 +0000 UTC m=+84.849708074 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.661693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.662319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.662342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.662368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.662388 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.765038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.765083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.765097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.765119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.765139 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.868086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.868148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.868163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.868187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.868203 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.971868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.971924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.971945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.971968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:17 crc kubenswrapper[4809]: I1127 17:10:17.971986 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:17Z","lastTransitionTime":"2025-11-27T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.075191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.075270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.075288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.075319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.075353 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:18Z","lastTransitionTime":"2025-11-27T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.178467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.178519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.178532 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.178553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.178570 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:18Z","lastTransitionTime":"2025-11-27T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.282159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.282212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.282225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.282247 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.282290 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:18Z","lastTransitionTime":"2025-11-27T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.385163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.385231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.385244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.385266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.385282 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:18Z","lastTransitionTime":"2025-11-27T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.457259 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:18 crc kubenswrapper[4809]: E1127 17:10:18.457484 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.489091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.489174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.489185 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.489204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.489215 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:18Z","lastTransitionTime":"2025-11-27T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.591960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.591994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.592004 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.592019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.592030 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:18Z","lastTransitionTime":"2025-11-27T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.694682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.694730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.694759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.694779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.694792 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:18Z","lastTransitionTime":"2025-11-27T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.798120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.798188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.798206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.798232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.798254 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:18Z","lastTransitionTime":"2025-11-27T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.901958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.902011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.902025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.902046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:18 crc kubenswrapper[4809]: I1127 17:10:18.902060 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:18Z","lastTransitionTime":"2025-11-27T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.005777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.005836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.005853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.005871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.005886 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.107942 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.107993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.108003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.108024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.108035 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.210173 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.210229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.210239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.210259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.210269 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.312385 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.312459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.312476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.312499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.312515 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.414536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.414613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.414625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.414648 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.414664 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.456881 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.456896 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:19 crc kubenswrapper[4809]: E1127 17:10:19.457000 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.457141 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:19 crc kubenswrapper[4809]: E1127 17:10:19.457198 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:19 crc kubenswrapper[4809]: E1127 17:10:19.457355 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.516629 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.516676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.516686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.516703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.516713 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.619628 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.619713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.619761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.619789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.619808 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.722136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.722218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.722234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.722258 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.722273 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.825821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.825896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.825937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.825981 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.826007 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.928184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.928245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.928261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.928284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:19 crc kubenswrapper[4809]: I1127 17:10:19.928297 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:19Z","lastTransitionTime":"2025-11-27T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.031061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.031134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.031170 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.031205 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.031228 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.135342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.135397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.135410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.135472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.135486 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.239471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.239549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.239570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.239600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.239622 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.297826 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.312028 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.315936 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.329255 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.342649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.342698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.342711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.342729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.342776 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.347225 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.365654 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.381336 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.399085 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.414121 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.426303 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.437934 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.446026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.446085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.446100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.446123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.446138 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.449962 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.457927 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:20 crc kubenswrapper[4809]: E1127 17:10:20.458174 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.464393 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.474871 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.489506 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.501122 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.517504 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.538565 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:20Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.548452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.548518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.548539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.548566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.548586 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.651352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.651402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.651413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.651431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.651442 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.754050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.754097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.754106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.754123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.754135 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.856077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.856121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.856130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.856145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.856155 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.958886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.958935 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.958947 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.958966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:20 crc kubenswrapper[4809]: I1127 17:10:20.959025 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:20Z","lastTransitionTime":"2025-11-27T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.061383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.061451 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.061466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.061484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.061495 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.164610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.164682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.164701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.164727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.164767 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.267584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.267614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.267625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.267642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.267653 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.371524 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.371587 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.371601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.371620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.371632 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.457921 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.457990 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.457966 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:21 crc kubenswrapper[4809]: E1127 17:10:21.458172 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:21 crc kubenswrapper[4809]: E1127 17:10:21.458300 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:21 crc kubenswrapper[4809]: E1127 17:10:21.458581 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.474507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.474575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.474595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.474621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.474659 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.580593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.580691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.580712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.580766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.580795 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.684907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.684968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.684980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.684996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.685007 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.787889 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.787957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.787995 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.788019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.788030 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.891765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.891822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.891832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.891851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.891863 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.994392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.994429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.994438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.994452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:21 crc kubenswrapper[4809]: I1127 17:10:21.994461 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:21Z","lastTransitionTime":"2025-11-27T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.097264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.097323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.097336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.097360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.097373 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:22Z","lastTransitionTime":"2025-11-27T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.200087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.200121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.200130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.200146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.200156 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:22Z","lastTransitionTime":"2025-11-27T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.303915 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.303972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.303989 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.304011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.304024 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:22Z","lastTransitionTime":"2025-11-27T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.406528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.406586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.406599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.406620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.406635 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:22Z","lastTransitionTime":"2025-11-27T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.456979 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:22 crc kubenswrapper[4809]: E1127 17:10:22.457196 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.510422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.510486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.510506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.510535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.510554 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:22Z","lastTransitionTime":"2025-11-27T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.614690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.614802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.614829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.614855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.614874 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:22Z","lastTransitionTime":"2025-11-27T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.718557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.718631 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.718654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.718688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.718711 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:22Z","lastTransitionTime":"2025-11-27T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.822904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.822978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.822998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.823025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.823043 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:22Z","lastTransitionTime":"2025-11-27T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.926207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.926271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.926280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.926297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:22 crc kubenswrapper[4809]: I1127 17:10:22.926307 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:22Z","lastTransitionTime":"2025-11-27T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.030310 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.030379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.030398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.030428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.030448 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.133428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.133506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.133527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.133554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.133573 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.237036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.237122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.237135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.237158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.237173 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.341215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.341287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.341302 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.341324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.341337 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.444226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.444308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.444334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.444370 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.444392 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.457704 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.457818 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.458076 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:23 crc kubenswrapper[4809]: E1127 17:10:23.458207 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:23 crc kubenswrapper[4809]: E1127 17:10:23.458434 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:23 crc kubenswrapper[4809]: E1127 17:10:23.458646 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.546599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.546667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.546686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.546714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.546732 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.650692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.650810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.650834 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.650862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.650881 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.754675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.754732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.754778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.754799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.754828 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.857091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.857150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.857166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.857190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.857208 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.960082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.960166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.960183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.960214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:23 crc kubenswrapper[4809]: I1127 17:10:23.960232 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:23Z","lastTransitionTime":"2025-11-27T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.063326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.063412 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.063441 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.063482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.063507 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.166443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.166508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.166521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.166537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.166549 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.270232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.270312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.270337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.270372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.270396 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.373614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.373662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.373696 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.373716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.373729 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.457362 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:24 crc kubenswrapper[4809]: E1127 17:10:24.457787 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.458622 4809 scope.go:117] "RemoveContainer" containerID="51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c" Nov 27 17:10:24 crc kubenswrapper[4809]: E1127 17:10:24.458876 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.478331 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.478508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.478537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.478616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.478644 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.581452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.581494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.581506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.581535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.581545 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.683960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.683999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.684008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.684022 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.684033 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.786981 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.787035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.787050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.787071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.787083 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.889833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.889885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.889894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.889911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.889924 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.993780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.993856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.993875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.993903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:24 crc kubenswrapper[4809]: I1127 17:10:24.993923 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:24Z","lastTransitionTime":"2025-11-27T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.096621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.096671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.096682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.096700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.096711 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.199486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.199578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.199598 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.199622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.199639 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.302523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.302570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.302578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.302600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.302611 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.405017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.405059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.405069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.405088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.405103 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.456880 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.456925 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.456945 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:25 crc kubenswrapper[4809]: E1127 17:10:25.457152 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:25 crc kubenswrapper[4809]: E1127 17:10:25.457280 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:25 crc kubenswrapper[4809]: E1127 17:10:25.457368 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.471236 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.483654 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.496808 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.508167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.508230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.508244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.508263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.508278 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.517346 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.531538 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.553172 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.577706 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.595291 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.611660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.611704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.611714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.611730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.611769 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.613234 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.626985 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.641063 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.653759 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.667786 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.684701 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.692088 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.699473 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.715081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.715146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.715159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.715180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.715225 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.721039 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.736473 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.757678 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.772226 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.786235 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.798854 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.815248 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.817144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.817188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.817198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.817216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.817228 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.827512 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.841523 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.854958 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.865771 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.880868 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.894880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.894932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.894943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.894958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.894969 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.899561 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: E1127 17:10:25.906472 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.910083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.910132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.910144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.910160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.910171 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.916846 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: E1127 17:10:25.924545 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.928494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.928553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.928569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.928589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.928601 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.930454 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: E1127 17:10:25.944086 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.946773 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.948443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.948475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.948486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.948505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.948517 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.959215 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: E1127 17:10:25.965689 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.969240 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.969284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.969293 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.969311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.969326 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.977271 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: E1127 17:10:25.983112 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:25 crc kubenswrapper[4809]: E1127 17:10:25.983233 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.985900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.985943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.985952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.985973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.985988 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:25Z","lastTransitionTime":"2025-11-27T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:25 crc kubenswrapper[4809]: I1127 17:10:25.990807 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:25Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.089078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.089131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.089146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.089166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.089180 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:26Z","lastTransitionTime":"2025-11-27T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.192102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.192144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.192157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.192175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.192189 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:26Z","lastTransitionTime":"2025-11-27T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.295791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.295872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.295899 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.295942 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.295969 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:26Z","lastTransitionTime":"2025-11-27T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.399574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.399649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.399667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.399698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.399718 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:26Z","lastTransitionTime":"2025-11-27T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.457113 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:26 crc kubenswrapper[4809]: E1127 17:10:26.457347 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.504488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.504549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.504566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.504608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.504628 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:26Z","lastTransitionTime":"2025-11-27T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.607274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.607558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.607704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.607826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.607978 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:26Z","lastTransitionTime":"2025-11-27T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.711954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.712020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.712035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.712059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.712074 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:26Z","lastTransitionTime":"2025-11-27T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.814659 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.814726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.814778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.814808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.814829 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:26Z","lastTransitionTime":"2025-11-27T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.918079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.918127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.918139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.918165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:26 crc kubenswrapper[4809]: I1127 17:10:26.918177 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:26Z","lastTransitionTime":"2025-11-27T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.021410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.021457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.021467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.021483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.021493 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.124177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.124215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.124226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.124239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.124251 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.227681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.227726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.227761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.227781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.227794 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.330866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.330918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.330929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.330952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.330967 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.435071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.435127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.435140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.435163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.435178 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.457392 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.457450 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.457566 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:27 crc kubenswrapper[4809]: E1127 17:10:27.457756 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:27 crc kubenswrapper[4809]: E1127 17:10:27.457844 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:27 crc kubenswrapper[4809]: E1127 17:10:27.457963 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.538490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.538549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.538565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.538589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.538605 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.642425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.642466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.642477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.642493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.642506 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.744908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.744950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.744961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.744985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.744997 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.847791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.847877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.847895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.848253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.848528 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.951855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.951903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.951913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.951931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:27 crc kubenswrapper[4809]: I1127 17:10:27.951946 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:27Z","lastTransitionTime":"2025-11-27T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.054870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.054939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.054950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.054964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.054973 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.157672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.157716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.157725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.157760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.157773 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.259568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.259599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.259606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.259644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.259655 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.362344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.362390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.362401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.362417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.362431 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.457538 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:28 crc kubenswrapper[4809]: E1127 17:10:28.457682 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.464779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.464829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.464841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.464860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.464873 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.566579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.566689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.566705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.566726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.566767 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.669803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.669921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.669956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.669995 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.670021 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.772726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.772785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.772795 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.772813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.772824 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.875488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.875537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.875547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.875563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.875574 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.979238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.979319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.979337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.979363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:28 crc kubenswrapper[4809]: I1127 17:10:28.979384 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:28Z","lastTransitionTime":"2025-11-27T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.082905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.082973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.082992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.083024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.083044 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:29Z","lastTransitionTime":"2025-11-27T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.186217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.186315 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.186332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.186357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.186376 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:29Z","lastTransitionTime":"2025-11-27T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.289489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.289538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.289548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.289566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.289576 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:29Z","lastTransitionTime":"2025-11-27T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.392920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.392961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.392978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.392995 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.393008 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:29Z","lastTransitionTime":"2025-11-27T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.457445 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.457566 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:29 crc kubenswrapper[4809]: E1127 17:10:29.457600 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.457669 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:29 crc kubenswrapper[4809]: E1127 17:10:29.457858 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:29 crc kubenswrapper[4809]: E1127 17:10:29.458862 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.495890 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.495932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.495942 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.495959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.495971 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:29Z","lastTransitionTime":"2025-11-27T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.598329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.598386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.598395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.598410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.598422 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:29Z","lastTransitionTime":"2025-11-27T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.704233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.704305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.704321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.704420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.704495 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:29Z","lastTransitionTime":"2025-11-27T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.806981 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.807019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.807030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.807047 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.807059 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:29Z","lastTransitionTime":"2025-11-27T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.909009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.909048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.909055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.909071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:29 crc kubenswrapper[4809]: I1127 17:10:29.909081 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:29Z","lastTransitionTime":"2025-11-27T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.011426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.011479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.011492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.011512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.011522 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.114965 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.115006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.115017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.115032 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.115042 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.218334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.218393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.218406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.218431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.218444 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.322447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.322538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.322592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.322618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.322677 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.426606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.426666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.426680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.426702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.426717 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.457450 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:30 crc kubenswrapper[4809]: E1127 17:10:30.457648 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.529676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.529761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.529774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.529794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.529806 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.632871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.632912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.632920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.632936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.632946 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.735287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.735354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.735373 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.735408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.735427 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.820412 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:30 crc kubenswrapper[4809]: E1127 17:10:30.820586 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:30 crc kubenswrapper[4809]: E1127 17:10:30.820655 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs podName:683f4d0e-67aa-40a7-b7b0-e5325ed0225f nodeName:}" failed. No retries permitted until 2025-11-27 17:11:02.820639619 +0000 UTC m=+98.093096971 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs") pod "network-metrics-daemon-2nc2d" (UID: "683f4d0e-67aa-40a7-b7b0-e5325ed0225f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.837949 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.837979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.837986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.838000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.838012 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.941150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.941486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.941570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.941683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:30 crc kubenswrapper[4809]: I1127 17:10:30.941774 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:30Z","lastTransitionTime":"2025-11-27T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.043618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.043651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.043661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.043674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.043684 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.146855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.146925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.146943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.146973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.147002 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.250343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.250384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.250396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.250414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.250425 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.353976 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.354028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.354044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.354063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.354076 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.456819 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.456948 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.456950 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:31 crc kubenswrapper[4809]: E1127 17:10:31.457139 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.457241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.457269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: E1127 17:10:31.457265 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.457283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.457349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: E1127 17:10:31.457369 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.457399 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.559672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.559804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.559823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.559849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.559868 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.662685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.662762 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.662777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.662794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.662805 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.765646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.765690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.765701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.765717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.765727 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.868846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.868887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.868896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.868914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.868924 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.971573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.971622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.971631 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.971647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:31 crc kubenswrapper[4809]: I1127 17:10:31.971661 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:31Z","lastTransitionTime":"2025-11-27T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.073880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.073925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.073972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.073993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.074005 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:32Z","lastTransitionTime":"2025-11-27T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.176478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.176529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.176541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.176561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.176574 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:32Z","lastTransitionTime":"2025-11-27T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.280805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.280839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.280858 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.280880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.280890 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:32Z","lastTransitionTime":"2025-11-27T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.384218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.384265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.384274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.384291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.384302 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:32Z","lastTransitionTime":"2025-11-27T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.457454 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:32 crc kubenswrapper[4809]: E1127 17:10:32.457631 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.487325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.487379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.487393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.487414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.487426 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:32Z","lastTransitionTime":"2025-11-27T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.590754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.590785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.590794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.590808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.590817 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:32Z","lastTransitionTime":"2025-11-27T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.693715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.693822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.693840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.693867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.693885 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:32Z","lastTransitionTime":"2025-11-27T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.797062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.797109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.797124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.797142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.797156 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:32Z","lastTransitionTime":"2025-11-27T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.899308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.899361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.899375 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.899398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:32 crc kubenswrapper[4809]: I1127 17:10:32.899419 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:32Z","lastTransitionTime":"2025-11-27T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.001493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.001531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.001565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.001585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.001595 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.104546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.104586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.104594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.104613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.104623 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.207211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.207260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.207272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.207293 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.207303 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.309159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.309202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.309213 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.309230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.309242 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.411987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.412032 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.412044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.412062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.412075 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.457251 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.457299 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.457281 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:33 crc kubenswrapper[4809]: E1127 17:10:33.457417 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:33 crc kubenswrapper[4809]: E1127 17:10:33.457522 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:33 crc kubenswrapper[4809]: E1127 17:10:33.457617 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.514778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.514826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.514835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.514849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.514865 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.617708 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.617823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.617837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.617859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.617872 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.720034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.720079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.720091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.720109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.720122 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.822935 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.822972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.822990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.823004 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.823015 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.926169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.926217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.926226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.926245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:33 crc kubenswrapper[4809]: I1127 17:10:33.926256 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:33Z","lastTransitionTime":"2025-11-27T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.029792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.029837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.029851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.029871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.029886 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.132981 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.133033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.133046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.133064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.133077 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.235835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.235886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.235900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.235927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.235940 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.338206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.338285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.338304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.338330 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.338373 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.441080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.441124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.441135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.441152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.441163 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.457391 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:34 crc kubenswrapper[4809]: E1127 17:10:34.457547 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.543994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.544044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.544055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.544071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.544082 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.647475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.647525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.647536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.647555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.647567 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.750365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.750405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.750417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.750434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.750446 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.852509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.852574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.852586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.852603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.852614 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.904994 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd8px_8741a6e8-a7f2-40ef-b25b-10e518345478/kube-multus/0.log" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.905057 4809 generic.go:334] "Generic (PLEG): container finished" podID="8741a6e8-a7f2-40ef-b25b-10e518345478" containerID="e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440" exitCode=1 Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.905091 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd8px" event={"ID":"8741a6e8-a7f2-40ef-b25b-10e518345478","Type":"ContainerDied","Data":"e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.909547 4809 scope.go:117] "RemoveContainer" containerID="e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.925222 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:34Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.943831 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:34Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.953978 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:34Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.954781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.954888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.954973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.955055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.955116 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:34Z","lastTransitionTime":"2025-11-27T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.966126 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:34Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.974673 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:34Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.985894 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:34Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:34 crc kubenswrapper[4809]: I1127 17:10:34.997523 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:34Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.011014 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.022894 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.038722 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.050715 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:33Z\\\",\\\"message\\\":\\\"2025-11-27T17:09:48+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6\\\\n2025-11-27T17:09:48+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6 to /host/opt/cni/bin/\\\\n2025-11-27T17:09:48Z [verbose] multus-daemon started\\\\n2025-11-27T17:09:48Z [verbose] Readiness Indicator file check\\\\n2025-11-27T17:10:33Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.058060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.058092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.058106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.058125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.058139 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.062694 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.073516 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.082055 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.092561 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.102344 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.114040 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.160623 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.160657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.160668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.160685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.160697 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.262689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.262755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.262770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.262787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.262799 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.365007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.365061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.365082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.365101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.365114 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.457366 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.457506 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.457513 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:35 crc kubenswrapper[4809]: E1127 17:10:35.457617 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:35 crc kubenswrapper[4809]: E1127 17:10:35.457719 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:35 crc kubenswrapper[4809]: E1127 17:10:35.457817 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.466930 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.466962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.466973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.466988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.466999 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.474505 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.493163 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.507840 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.518427 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.527647 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.539925 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.551585 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.568799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.568839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.568853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.568873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.568884 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.571306 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.583195 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.597053 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.611285 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:33Z\\\",\\\"message\\\":\\\"2025-11-27T17:09:48+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6\\\\n2025-11-27T17:09:48+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6 to /host/opt/cni/bin/\\\\n2025-11-27T17:09:48Z [verbose] multus-daemon started\\\\n2025-11-27T17:09:48Z [verbose] Readiness Indicator file check\\\\n2025-11-27T17:10:33Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.631255 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.644910 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.655949 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.669028 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.670996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.671039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.671047 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.671064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.671074 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.680701 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.690567 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.772896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.773206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.773312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.773411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.773504 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.875825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.875863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.875872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.875888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.875897 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.910296 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd8px_8741a6e8-a7f2-40ef-b25b-10e518345478/kube-multus/0.log" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.910362 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd8px" event={"ID":"8741a6e8-a7f2-40ef-b25b-10e518345478","Type":"ContainerStarted","Data":"aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.923597 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.941884 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.954803 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.969657 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.978445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.978482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.978491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.978506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.978517 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:35Z","lastTransitionTime":"2025-11-27T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.981462 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:35 crc kubenswrapper[4809]: I1127 17:10:35.994696 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:35Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.007504 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.022770 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.037369 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.049917 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.066455 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:33Z\\\",\\\"message\\\":\\\"2025-11-27T17:09:48+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6\\\\n2025-11-27T17:09:48+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6 to /host/opt/cni/bin/\\\\n2025-11-27T17:09:48Z [verbose] multus-daemon started\\\\n2025-11-27T17:09:48Z [verbose] Readiness Indicator file check\\\\n2025-11-27T17:10:33Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.079067 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.080952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.080981 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.080997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.081019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.081029 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.091897 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.104288 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.113443 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.126614 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.138294 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.174806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.174852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.174861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.174879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.174892 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: E1127 17:10:36.187921 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.191118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.191154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.191165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.191181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.191191 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: E1127 17:10:36.203965 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.210936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.210993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.211011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.211038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.211055 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: E1127 17:10:36.222629 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.226342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.226378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.226387 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.226406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.226418 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: E1127 17:10:36.238222 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.243057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.243337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.243488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.243599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.243680 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: E1127 17:10:36.254700 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: E1127 17:10:36.255057 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.256841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.256956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.257038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.257113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.257188 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.359341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.359383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.359394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.359410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.359422 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.457785 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:36 crc kubenswrapper[4809]: E1127 17:10:36.458244 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.458588 4809 scope.go:117] "RemoveContainer" containerID="51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.461072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.461106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.461119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.461136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.461148 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.564291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.564397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.564411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.564554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.564587 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.667671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.667710 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.667719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.667753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.667764 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.770493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.770547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.770561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.770581 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.770595 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.873591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.873651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.873664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.873688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.873701 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.916614 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/2.log" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.919586 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.920124 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.966697 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.976922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.976969 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.976980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.977002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.977015 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:36Z","lastTransitionTime":"2025-11-27T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:36 crc kubenswrapper[4809]: I1127 17:10:36.994964 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:36Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.030395 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.048816 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.063819 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.078658 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.079198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.079288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.079304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.079326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.079338 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:37Z","lastTransitionTime":"2025-11-27T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.096678 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.112461 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.128088 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:33Z\\\",\\\"message\\\":\\\"2025-11-27T17:09:48+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6\\\\n2025-11-27T17:09:48+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6 to /host/opt/cni/bin/\\\\n2025-11-27T17:09:48Z [verbose] multus-daemon started\\\\n2025-11-27T17:09:48Z [verbose] Readiness Indicator file check\\\\n2025-11-27T17:10:33Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.143697 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.160512 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.176551 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.185458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.185563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.185627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.185655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.185693 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:37Z","lastTransitionTime":"2025-11-27T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.193412 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.207286 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.223564 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.238223 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.253655 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.289197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.289267 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.289278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.289298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.289310 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:37Z","lastTransitionTime":"2025-11-27T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.392667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.392781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.392797 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.392822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.392835 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:37Z","lastTransitionTime":"2025-11-27T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.458285 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:37 crc kubenswrapper[4809]: E1127 17:10:37.458445 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.458655 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:37 crc kubenswrapper[4809]: E1127 17:10:37.458701 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.458887 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:37 crc kubenswrapper[4809]: E1127 17:10:37.459059 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.495960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.496004 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.496014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.496032 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.496043 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:37Z","lastTransitionTime":"2025-11-27T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.600368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.600437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.600448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.600467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.600480 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:37Z","lastTransitionTime":"2025-11-27T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.703006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.703057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.703067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.703084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.703099 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:37Z","lastTransitionTime":"2025-11-27T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.806625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.806700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.806718 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.806859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.806879 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:37Z","lastTransitionTime":"2025-11-27T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.910284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.910337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.910353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.910379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.910390 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:37Z","lastTransitionTime":"2025-11-27T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.926591 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/3.log" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.927672 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/2.log" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.932975 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" exitCode=1 Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.933061 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d"} Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.933113 4809 scope.go:117] "RemoveContainer" containerID="51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.934566 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:10:37 crc kubenswrapper[4809]: E1127 17:10:37.934865 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.947610 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.965448 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:37 crc kubenswrapper[4809]: I1127 17:10:37.991375 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51ecabddcf80b8c16fbf197359322f01727157b1986351a30a3bf1c50c25ed6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:12Z\\\",\\\"message\\\":\\\"etwork controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:12Z is after 2025-08-24T17:21:41Z]\\\\nI1127 17:10:12.345522 6474 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-daemon\\\\\\\"}\\\\nI1127 17:10:12.345515 6474 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fiel\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:37Z\\\",\\\"message\\\":\\\" 6835 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 17:10:37.438821 6835 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-qwx9w\\\\nI1127 17:10:37.438828 6835 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1127 17:10:37.438833 6835 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-qwx9w\\\\nF1127 17:10:37.438829 6835 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.007404 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.012534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.012592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.012603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.012625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.012637 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.020536 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.037984 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.053897 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:33Z\\\",\\\"message\\\":\\\"2025-11-27T17:09:48+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6\\\\n2025-11-27T17:09:48+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6 to /host/opt/cni/bin/\\\\n2025-11-27T17:09:48Z [verbose] multus-daemon started\\\\n2025-11-27T17:09:48Z [verbose] Readiness Indicator file check\\\\n2025-11-27T17:10:33Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.068261 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.086488 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.105804 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.115629 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.115880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.115949 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.116028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.116112 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.121341 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.136363 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.153081 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.165228 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.181296 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.192632 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.213125 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.219300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.219330 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.219340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.219379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.219394 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.322701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.322885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.322912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.322941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.322957 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.425816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.425867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.425878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.425895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.425906 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.457265 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:38 crc kubenswrapper[4809]: E1127 17:10:38.458863 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.529083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.529135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.529147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.529169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.529184 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.632506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.632601 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.632627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.632660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.632687 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.735965 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.736028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.736041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.736063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.736078 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.840298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.840353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.840367 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.840387 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.840401 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.939176 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/3.log" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.942114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.942148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.942159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.942174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.942186 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:38Z","lastTransitionTime":"2025-11-27T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.942512 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:10:38 crc kubenswrapper[4809]: E1127 17:10:38.942654 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.956989 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.981208 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:38 crc kubenswrapper[4809]: I1127 17:10:38.996098 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:38Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.012756 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.028399 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:33Z\\\",\\\"message\\\":\\\"2025-11-27T17:09:48+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6\\\\n2025-11-27T17:09:48+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6 to /host/opt/cni/bin/\\\\n2025-11-27T17:09:48Z [verbose] multus-daemon started\\\\n2025-11-27T17:09:48Z [verbose] Readiness Indicator file check\\\\n2025-11-27T17:10:33Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.042202 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.044794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.044836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.044847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.044866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.044880 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.057279 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.068759 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.080866 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.091546 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.101578 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.115949 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.138476 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:37Z\\\",\\\"message\\\":\\\" 6835 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 17:10:37.438821 6835 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-qwx9w\\\\nI1127 17:10:37.438828 6835 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1127 17:10:37.438833 6835 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-qwx9w\\\\nF1127 17:10:37.438829 6835 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.148010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.148048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.148062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.148082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.148095 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.151943 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.165601 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.176182 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.191910 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:39Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.250934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.250984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.250997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.251016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.251027 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.355099 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.355195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.355215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.355251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.355278 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.457079 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.457307 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:39 crc kubenswrapper[4809]: E1127 17:10:39.457491 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.457556 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:39 crc kubenswrapper[4809]: E1127 17:10:39.457699 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:39 crc kubenswrapper[4809]: E1127 17:10:39.457887 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.458792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.458837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.458856 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.458879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.458902 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.562584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.562658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.562673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.562695 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.562709 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.666081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.666155 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.666173 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.666200 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.666227 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.770145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.770210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.770225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.770249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.770265 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.873612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.873682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.873701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.873733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.873776 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.977580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.977631 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.977641 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.977661 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:39 crc kubenswrapper[4809]: I1127 17:10:39.977674 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:39Z","lastTransitionTime":"2025-11-27T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.080907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.080964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.080978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.081003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.081019 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:40Z","lastTransitionTime":"2025-11-27T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.184408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.184530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.184551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.184596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.184620 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:40Z","lastTransitionTime":"2025-11-27T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.288945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.289012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.289030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.289057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.289075 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:40Z","lastTransitionTime":"2025-11-27T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.392226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.392309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.392333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.392363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.392382 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:40Z","lastTransitionTime":"2025-11-27T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.457337 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:40 crc kubenswrapper[4809]: E1127 17:10:40.457552 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.495068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.495115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.495129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.495147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.495160 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:40Z","lastTransitionTime":"2025-11-27T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.598253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.598319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.598332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.598352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.598368 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:40Z","lastTransitionTime":"2025-11-27T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.701988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.702040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.702055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.702075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.702092 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:40Z","lastTransitionTime":"2025-11-27T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.805265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.805316 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.805332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.805356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.805371 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:40Z","lastTransitionTime":"2025-11-27T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.908995 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.909386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.909449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.909513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:40 crc kubenswrapper[4809]: I1127 17:10:40.909583 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:40Z","lastTransitionTime":"2025-11-27T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.013162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.013221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.013236 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.013260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.013280 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.117010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.117072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.117086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.117108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.117123 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.220696 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.221127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.221196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.221261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.221318 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.324689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.325120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.325234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.325354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.325463 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.428876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.428931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.428944 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.428968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.428982 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.457709 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.457816 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.457816 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:41 crc kubenswrapper[4809]: E1127 17:10:41.457969 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:41 crc kubenswrapper[4809]: E1127 17:10:41.458110 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:41 crc kubenswrapper[4809]: E1127 17:10:41.458209 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.532238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.532325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.532340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.532361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.532377 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.636277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.636431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.636450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.636472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.636486 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.740055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.740123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.740146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.740174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.740195 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.843314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.843361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.843375 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.843421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.843437 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.946101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.946154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.946167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.946186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:41 crc kubenswrapper[4809]: I1127 17:10:41.946199 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:41Z","lastTransitionTime":"2025-11-27T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.048943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.048998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.049009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.049029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.049043 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.152252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.152309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.152323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.152348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.152362 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.255131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.255191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.255208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.255233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.255247 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.358630 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.358685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.358701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.358724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.358761 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.456814 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:42 crc kubenswrapper[4809]: E1127 17:10:42.457001 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.461961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.461993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.462001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.462014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.462025 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.565764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.565819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.565830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.565847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.565859 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.668919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.668968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.668980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.669001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.669013 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.777902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.777973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.777992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.778018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.778038 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.880688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.880756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.880774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.880796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.880811 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.984006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.984060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.984076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.984098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:42 crc kubenswrapper[4809]: I1127 17:10:42.984113 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:42Z","lastTransitionTime":"2025-11-27T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.087276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.087343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.087361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.087386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.087405 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:43Z","lastTransitionTime":"2025-11-27T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.190167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.190215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.190224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.190241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.190251 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:43Z","lastTransitionTime":"2025-11-27T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.293621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.293673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.293683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.293703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.293717 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:43Z","lastTransitionTime":"2025-11-27T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.396311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.396348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.396358 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.396372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.396381 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:43Z","lastTransitionTime":"2025-11-27T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.457480 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.457554 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:43 crc kubenswrapper[4809]: E1127 17:10:43.457690 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.457725 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:43 crc kubenswrapper[4809]: E1127 17:10:43.457799 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:43 crc kubenswrapper[4809]: E1127 17:10:43.457951 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.499061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.499103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.499113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.499133 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.499145 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:43Z","lastTransitionTime":"2025-11-27T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.602190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.602230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.602239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.602255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.602268 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:43Z","lastTransitionTime":"2025-11-27T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.705033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.705099 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.705111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.705146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.705159 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:43Z","lastTransitionTime":"2025-11-27T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.807818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.807863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.807872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.807893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.807903 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:43Z","lastTransitionTime":"2025-11-27T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.910127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.910191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.910203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.910223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:43 crc kubenswrapper[4809]: I1127 17:10:43.910237 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:43Z","lastTransitionTime":"2025-11-27T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.013123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.013202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.013241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.013261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.013274 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.116159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.116206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.116217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.116234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.116247 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.219252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.219316 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.219338 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.219368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.219391 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.322704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.322815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.322835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.322860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.322879 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.425412 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.425482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.425500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.425530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.425552 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.457874 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:44 crc kubenswrapper[4809]: E1127 17:10:44.458050 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.528637 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.528716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.528764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.528798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.528823 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.631491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.631589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.631612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.631636 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.631654 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.735337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.735380 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.735393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.735411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.735422 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.838082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.838129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.838143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.838165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.838183 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.941272 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.941345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.941372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.941405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:44 crc kubenswrapper[4809]: I1127 17:10:44.941429 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:44Z","lastTransitionTime":"2025-11-27T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.044969 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.045046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.045071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.045104 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.045128 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.147592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.147638 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.147651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.147671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.147684 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.250425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.250503 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.250522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.250550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.250582 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.353273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.353346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.353360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.353384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.353405 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.455425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.455477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.455491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.455511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.455527 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.456887 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.456935 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.456967 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:45 crc kubenswrapper[4809]: E1127 17:10:45.457065 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:45 crc kubenswrapper[4809]: E1127 17:10:45.457159 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:45 crc kubenswrapper[4809]: E1127 17:10:45.457255 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.472253 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.491300 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:37Z\\\",\\\"message\\\":\\\" 6835 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 17:10:37.438821 6835 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-qwx9w\\\\nI1127 17:10:37.438828 6835 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1127 17:10:37.438833 6835 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-qwx9w\\\\nF1127 17:10:37.438829 6835 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.502650 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.512986 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.524039 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.536033 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.549932 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.557232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.557268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.557277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.557294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.557305 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.564449 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.577205 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.590968 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:33Z\\\",\\\"message\\\":\\\"2025-11-27T17:09:48+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6\\\\n2025-11-27T17:09:48+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6 to /host/opt/cni/bin/\\\\n2025-11-27T17:09:48Z [verbose] multus-daemon started\\\\n2025-11-27T17:09:48Z [verbose] Readiness Indicator file check\\\\n2025-11-27T17:10:33Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.606808 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.619796 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.635017 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.649029 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.659383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.659428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.659445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.659465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.659477 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.659953 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.671470 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.682991 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:45Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.761704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.761952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.761962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.761978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.761988 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.864316 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.864356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.864366 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.864383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.864396 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.967829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.967893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.967910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.967934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:45 crc kubenswrapper[4809]: I1127 17:10:45.967949 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:45Z","lastTransitionTime":"2025-11-27T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.069983 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.070042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.070059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.070080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.070092 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.172078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.172116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.172129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.172145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.172157 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.274722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.274799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.274809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.274827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.274838 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.378088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.378166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.378183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.378204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.378219 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.457563 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:46 crc kubenswrapper[4809]: E1127 17:10:46.457733 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.462151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.462215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.462232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.462253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.462265 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: E1127 17:10:46.475683 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:46Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.480132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.480179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.480192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.480211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.480227 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: E1127 17:10:46.501125 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:46Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.507045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.507109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.507135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.507156 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.507170 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: E1127 17:10:46.521965 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:46Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.526101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.526140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.526151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.526172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.526185 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: E1127 17:10:46.539171 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:46Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.543900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.543936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.543948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.543962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.543974 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: E1127 17:10:46.556983 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:46Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:46 crc kubenswrapper[4809]: E1127 17:10:46.557107 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.558555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.558638 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.558694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.558725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.558775 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.662452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.662511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.662522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.662545 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.662561 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.766317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.766397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.766410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.766437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.766451 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.869970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.870014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.870024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.870042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.870052 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.971576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.971626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.971640 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.971657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:46 crc kubenswrapper[4809]: I1127 17:10:46.971669 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:46Z","lastTransitionTime":"2025-11-27T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.074789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.074860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.074873 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.074899 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.074913 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:47Z","lastTransitionTime":"2025-11-27T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.178403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.178463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.178474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.178492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.178504 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:47Z","lastTransitionTime":"2025-11-27T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.281713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.281778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.281792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.281809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.281820 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:47Z","lastTransitionTime":"2025-11-27T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.384733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.384804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.384818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.384837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.384852 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:47Z","lastTransitionTime":"2025-11-27T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.457304 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.457317 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.457320 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:47 crc kubenswrapper[4809]: E1127 17:10:47.457589 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:47 crc kubenswrapper[4809]: E1127 17:10:47.457460 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:47 crc kubenswrapper[4809]: E1127 17:10:47.457638 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.487221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.487251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.487259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.487270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.487279 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:47Z","lastTransitionTime":"2025-11-27T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.589526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.589565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.589576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.589591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.589601 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:47Z","lastTransitionTime":"2025-11-27T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.692018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.692062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.692070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.692089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.692100 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:47Z","lastTransitionTime":"2025-11-27T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.794714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.794790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.794803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.794823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.794838 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:47Z","lastTransitionTime":"2025-11-27T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.897792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.897846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.897859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.897877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:47 crc kubenswrapper[4809]: I1127 17:10:47.897889 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:47Z","lastTransitionTime":"2025-11-27T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.000556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.000603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.000618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.000640 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.000655 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.104401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.104561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.104587 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.104614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.104635 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.207478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.207530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.207543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.207563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.207576 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.309911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.310001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.310015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.310033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.310045 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.412962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.413000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.413010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.413023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.413033 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.456848 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:48 crc kubenswrapper[4809]: E1127 17:10:48.456993 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.515123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.515192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.515210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.515233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.515251 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.618256 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.618320 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.618337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.618363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.618383 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.721040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.721094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.721110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.721130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.721145 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.823939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.823978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.823991 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.824035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.824048 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.926170 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.926264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.926279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.926300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:48 crc kubenswrapper[4809]: I1127 17:10:48.926313 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:48Z","lastTransitionTime":"2025-11-27T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.028564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.028597 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.028606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.028620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.028630 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.130881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.130944 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.130957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.130974 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.130986 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.233134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.233172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.233184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.233202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.233216 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.335785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.335848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.335858 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.335874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.335885 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.438234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.438280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.438290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.438306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.438350 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.457474 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.457603 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.457497 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.457883 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.458229 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.458387 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.458564 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.458793 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.472007 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.529434 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.529599 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:53.529569059 +0000 UTC m=+148.802026421 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.529672 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.529785 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.529813 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.529885 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.529935 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:11:53.529919578 +0000 UTC m=+148.802376920 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.529976 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.530001 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.530015 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.530012 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.530056 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 17:11:53.530045683 +0000 UTC m=+148.802503165 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.530104 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 17:11:53.530083814 +0000 UTC m=+148.802541166 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.540618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.540659 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.540673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.540690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.540700 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.630952 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.631106 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.631121 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.631132 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:49 crc kubenswrapper[4809]: E1127 17:10:49.631214 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 17:11:53.631199203 +0000 UTC m=+148.903656555 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.643570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.643624 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.643637 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.643659 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.643672 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.747425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.747469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.747478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.747494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.747503 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.850079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.850128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.850140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.850160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.850172 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.952816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.952881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.952898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.952916 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:49 crc kubenswrapper[4809]: I1127 17:10:49.952928 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:49Z","lastTransitionTime":"2025-11-27T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.056266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.056658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.056670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.056689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.056704 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.158813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.158894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.158923 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.158965 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.158988 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.261356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.261410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.261427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.261458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.261483 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.363273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.363315 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.363327 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.363344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.363353 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.457560 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:50 crc kubenswrapper[4809]: E1127 17:10:50.457715 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.465227 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.465257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.465267 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.465282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.465294 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.568049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.568121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.568144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.568174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.568196 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.670722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.670785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.670797 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.670812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.670824 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.773579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.773618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.773627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.773642 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.773653 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.876606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.876645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.876656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.876672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.876686 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.979255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.979319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.979361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.979390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:50 crc kubenswrapper[4809]: I1127 17:10:50.979410 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:50Z","lastTransitionTime":"2025-11-27T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.081959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.082030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.082053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.082081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.082104 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:51Z","lastTransitionTime":"2025-11-27T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.185037 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.185087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.185096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.185112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.185121 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:51Z","lastTransitionTime":"2025-11-27T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.287527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.287576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.287599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.287620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.287634 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:51Z","lastTransitionTime":"2025-11-27T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.390500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.390550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.390559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.390574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.390583 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:51Z","lastTransitionTime":"2025-11-27T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.457012 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:51 crc kubenswrapper[4809]: E1127 17:10:51.457438 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.457143 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:51 crc kubenswrapper[4809]: E1127 17:10:51.457693 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.457128 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:51 crc kubenswrapper[4809]: E1127 17:10:51.457987 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.492405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.492460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.492474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.492492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.492506 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:51Z","lastTransitionTime":"2025-11-27T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.595665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.595723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.595745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.595758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.595768 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:51Z","lastTransitionTime":"2025-11-27T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.697933 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.697961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.697970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.697983 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.697992 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:51Z","lastTransitionTime":"2025-11-27T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.800199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.800243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.800251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.800268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.800279 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:51Z","lastTransitionTime":"2025-11-27T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.902985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.903030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.903038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.903053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:51 crc kubenswrapper[4809]: I1127 17:10:51.903063 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:51Z","lastTransitionTime":"2025-11-27T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.005362 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.005405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.005420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.005438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.005452 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.107556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.107660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.107678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.107697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.107707 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.209683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.209775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.209793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.209821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.209840 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.313288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.313540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.313603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.313715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.313823 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.416544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.416934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.417016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.417097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.417189 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.457042 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:52 crc kubenswrapper[4809]: E1127 17:10:52.457203 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.519488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.519572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.519584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.519604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.519616 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.622372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.622443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.622468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.622499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.622523 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.725618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.725672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.725684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.725702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.725715 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.828034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.828085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.828098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.828154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.828168 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.930646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.930691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.930704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.930721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:52 crc kubenswrapper[4809]: I1127 17:10:52.930732 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:52Z","lastTransitionTime":"2025-11-27T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.033662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.033714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.033724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.033757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.033769 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.135907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.135944 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.135953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.135968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.135977 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.238971 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.239052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.239074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.239108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.239132 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.340970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.341008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.341018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.341053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.341068 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.443475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.443546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.443564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.443590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.443611 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.457903 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.457968 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.458151 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:53 crc kubenswrapper[4809]: E1127 17:10:53.458261 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:53 crc kubenswrapper[4809]: E1127 17:10:53.458396 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:53 crc kubenswrapper[4809]: E1127 17:10:53.458586 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.545931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.546093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.546124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.546157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.546181 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.649217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.649270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.649311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.649334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.649352 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.753441 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.753538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.753570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.753605 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.753631 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.856363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.856999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.857046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.857077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.857098 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.959989 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.960074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.960105 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.960138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:53 crc kubenswrapper[4809]: I1127 17:10:53.960160 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:53Z","lastTransitionTime":"2025-11-27T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.063255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.063299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.063309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.063328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.063340 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.166231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.166304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.166324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.166346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.166364 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.269233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.269288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.269301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.269321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.269335 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.372046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.372083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.372091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.372106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.372117 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.457217 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:54 crc kubenswrapper[4809]: E1127 17:10:54.457387 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.475021 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.475067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.475111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.475128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.475150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.475167 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.578436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.578501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.578518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.578547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.578565 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.681668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.681731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.681777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.681798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.681813 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.784539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.784600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.784611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.784629 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.784640 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.887323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.887381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.887390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.887407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.887420 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.990341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.990382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.990392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.990409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:54 crc kubenswrapper[4809]: I1127 17:10:54.990421 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:54Z","lastTransitionTime":"2025-11-27T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.093416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.093473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.093494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.093527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.093554 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:55Z","lastTransitionTime":"2025-11-27T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.196702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.196778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.196789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.196808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.196821 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:55Z","lastTransitionTime":"2025-11-27T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.299310 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.299353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.299364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.299382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.299397 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:55Z","lastTransitionTime":"2025-11-27T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.401535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.401580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.401592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.401608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.401620 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:55Z","lastTransitionTime":"2025-11-27T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.457552 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.457627 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:55 crc kubenswrapper[4809]: E1127 17:10:55.457700 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.457795 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:55 crc kubenswrapper[4809]: E1127 17:10:55.457901 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:55 crc kubenswrapper[4809]: E1127 17:10:55.457945 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.471143 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69905841-f2e5-433a-a3f4-86114110cb22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6516e0de7b28debd9e9fc35e40f4ecadcd6388021cd8642f9540e6ea1aea7fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d508b1c13d6690a9dab7dfedf8148a92bf7e58d25c8daeb42cc379c29933fb08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d1f4e1caed59793e115c07eb79919ab9bf75510232b3d575621c74a1edb056a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.481733 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fr7qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdbfe7ca-6631-4c54-a71a-4b73cae5a64a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93c47d4ee41274a6eec64ad4b95184f6e120faec3c09c9aa23aedb1f35513e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8bqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fr7qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.493250 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ff0184d4c30bffdb3b61d2fdf4e67e539710a256861485cec5e1b08bdfdb0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.503817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.503860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.503871 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.503890 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.503900 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:55Z","lastTransitionTime":"2025-11-27T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.507494 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.520414 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21848912-a128-4704-b84a-811af21e99cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31ed6a42e19f8b28870b2e5e276a9a89e0cd65c8186106b8c4ad6cb65c79bcfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209c6e09a83af0fb6ae057bbf5815f9f98215d213333a2df9b04e24fb4494aa5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ca1353f3c8d04cbba38b389b7de4afa2dde7ee805d7d2dbb90b98ac9c2c7392\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d17bbd2e4e4ca8857f5acf18daf5d40b7dd162e044c98b0347d555db440f6121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8fbf1a50ce2a2ccf46693692a1e9c2b7bbe093fb5693df0be7cc78622decf64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cb92a62f68e762c1707e8e88c87c01a6415e48df3ae0339cb1b92264c1de4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede4204ea3a68ac9ca8814f8f42d4a9b168672f974dd26ee31b9e37afd511c53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2jxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4g6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.532285 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.543650 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.557935 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd8px" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8741a6e8-a7f2-40ef-b25b-10e518345478\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:33Z\\\",\\\"message\\\":\\\"2025-11-27T17:09:48+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6\\\\n2025-11-27T17:09:48+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_af6f36a5-068b-41c5-9662-0e1a408416f6 to /host/opt/cni/bin/\\\\n2025-11-27T17:09:48Z [verbose] multus-daemon started\\\\n2025-11-27T17:09:48Z [verbose] Readiness Indicator file check\\\\n2025-11-27T17:10:33Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-529wf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd8px\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.572165 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54849f6c-4354-465d-85cc-f118c0ca795c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbd8401ad7e86e42b870c9bd8f7849a31f2e1df23601d207492ce2a4e80520ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5793c61a7a3442194329d2d7ea7e68ea107b4804b08d90f837b0ce8f7a4beff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0ce223ada6fe1b30c1219173e980463e3ff9dedf67f81785c46de760cb4521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6013a3ba12d72dcdb6c9db2b68e17efaec0963d4d4fa84429321b78ef758a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.587532 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f50ffb44d12e6728b0d2065f4aff3d197304552c2f70d05e94fc5bda4dc2199e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bd901784f5a6ed2279b8840ec92b5790682ee4c872f4746cc4fa5478d90f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.597981 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-d6cbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ce6b2e-1bc1-438a-8c0d-e6b6e0a2bf2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9ca20e66ac06eda372856fd1dd5c480e962707cc8144344aa7749c8454fe2eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgzf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-d6cbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.606509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.606548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.606557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.606572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.606581 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:55Z","lastTransitionTime":"2025-11-27T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.610185 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2081b073677d38774097430154e0afca5c73bfe612a26b196c17e05fbe62c4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.619441 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afa276bf-d508-44a3-bde0-c9a96a4de65d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c37eb726c28c6b01b98224633c573380f722f84af8f66509fa1e053f8dce025\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4db2d583ba0e4df7c84a0866bc811893a201b8ab487a2b34b98447353a48373f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwzt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pt22c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.628837 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79cvs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nc2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.650851 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b549f7f-58e1-47de-831c-0d6e958f5440\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://168aba809229b0e356f02151ec9fc48f2aafa1de882b9a080a17f36f38570c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34401cef59489038f8fb148eb3db776826ca6ce79a3428846d1b3f3d102e71ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9c71189531579749b27b0c49e83d98646a4222aefc06e29b3f7822230dd25ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61f28e0a9ca53636faae3bddfe1346a9bb416102145e38fff56ad1fd0d06f735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://847769dbc77952b4382055d8b6aacd5a6cfde6d3e5b5c6f3d2262698929c15d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5628dbf8173e6827b814898bfdd58d1df9e0ae22939e3756a3d1f71b8f7a123d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5628dbf8173e6827b814898bfdd58d1df9e0ae22939e3756a3d1f71b8f7a123d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07ee0db804d52fa1651d5eee7bf0a1ddb81bf96e5a616cc5e7b915c149996901\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07ee0db804d52fa1651d5eee7bf0a1ddb81bf96e5a616cc5e7b915c149996901\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f4824645026999e711ed4c5dc7f8857a33e0bfeac58cf6e77d798a9f9ab699a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4824645026999e711ed4c5dc7f8857a33e0bfeac58cf6e77d798a9f9ab699a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.666012 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"518cb4a7-b48a-42fc-9436-13b6e54f9697\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"W1127 17:09:39.848370 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 17:09:39.848828 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764263379 cert, and key in /tmp/serving-cert-3394512977/serving-signer.crt, /tmp/serving-cert-3394512977/serving-signer.key\\\\nI1127 17:09:40.235118 1 observer_polling.go:159] Starting file observer\\\\nW1127 17:09:45.248203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 17:09:45.248435 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 17:09:45.249406 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3394512977/tls.crt::/tmp/serving-cert-3394512977/tls.key\\\\\\\"\\\\nI1127 17:09:45.596474 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 17:09:45.602941 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 17:09:45.602972 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 17:09:45.603003 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 17:09:45.603010 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 17:09:45.626695 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 17:09:45.626726 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1127 17:09:45.627806 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 17:09:45.630070 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.677071 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb082338-a0dd-47b1-ab5b-fa5d2d553d1a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1be54422827bedc61e9f1a13630540aac14ef91a65d85dfea89dcc578e0d90c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725750eead54ee1b84c7ec6c15e16019f58e9bfdc099f28d286076b2907f8235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://725750eead54ee1b84c7ec6c15e16019f58e9bfdc099f28d286076b2907f8235\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.697932 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b4318a-f089-451d-8a16-97de26acce28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T17:10:37Z\\\",\\\"message\\\":\\\" 6835 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 17:10:37.438821 6835 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-qwx9w\\\\nI1127 17:10:37.438828 6835 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1127 17:10:37.438833 6835 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-qwx9w\\\\nF1127 17:10:37.438829 6835 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:37Z is after 2025-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T17:10:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-52b7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m5b98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.708798 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b63ded2-aa4f-4aba-b3cc-a3965a01036c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T17:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6626b4c2f9505a8596334099e9b5545af1af937baf7dea77e7877c0b83d4b83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T17:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pdhtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T17:09:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qwx9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:55Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.709001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.709026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.709037 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.709052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.709062 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:55Z","lastTransitionTime":"2025-11-27T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.811337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.811682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.811799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.811903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.811968 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:55Z","lastTransitionTime":"2025-11-27T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.913868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.913914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.913928 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.913948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:55 crc kubenswrapper[4809]: I1127 17:10:55.913961 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:55Z","lastTransitionTime":"2025-11-27T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.015612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.015651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.015660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.015676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.015688 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.117905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.117936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.117944 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.117959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.117968 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.221018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.221057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.221068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.221085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.221095 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.323993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.324028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.324036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.324051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.324063 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.426730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.426877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.426895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.426921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.426941 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.457444 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:56 crc kubenswrapper[4809]: E1127 17:10:56.457625 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.529291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.529323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.529330 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.529343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.529352 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.632264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.632352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.632376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.632415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.632435 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.735317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.735391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.735405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.735429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.735444 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.838775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.838818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.838827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.838843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.838855 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.941500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.941546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.941558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.941575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.941586 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.957302 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.957338 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.957497 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.957515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.957527 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: E1127 17:10:56.975925 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.979479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.979516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.979525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.979541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.979552 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:56 crc kubenswrapper[4809]: E1127 17:10:56.991910 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:56Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.995658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.995720 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.995768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.995804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:56 crc kubenswrapper[4809]: I1127 17:10:56.995825 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:56Z","lastTransitionTime":"2025-11-27T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: E1127 17:10:57.010668 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.014084 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.014129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.014142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.014164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.014181 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: E1127 17:10:57.026349 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.029693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.029753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.029764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.029782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.029797 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: E1127 17:10:57.042234 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T17:10:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37894acd-0695-4c66-9d28-ea598b1264ef\\\",\\\"systemUUID\\\":\\\"2fa873ac-d151-4e97-8902-297a67778c76\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T17:10:57Z is after 2025-08-24T17:21:41Z" Nov 27 17:10:57 crc kubenswrapper[4809]: E1127 17:10:57.042351 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.043922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.043966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.043977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.043995 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.044009 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.147364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.147417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.147427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.147447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.147458 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.250083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.250119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.250127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.250158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.250171 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.352852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.352891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.352902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.352925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.352941 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.456799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.456885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.456910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.456945 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.456970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.456988 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.456885 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:57 crc kubenswrapper[4809]: E1127 17:10:57.457114 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.456907 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:57 crc kubenswrapper[4809]: E1127 17:10:57.457274 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:57 crc kubenswrapper[4809]: E1127 17:10:57.457443 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.560171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.560223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.560235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.560253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.560269 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.662578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.662658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.662683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.662712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.662732 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.765835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.765897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.765910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.765927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.765941 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.868058 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.868121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.868134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.868159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.868172 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.970832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.970911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.970932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.970957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:57 crc kubenswrapper[4809]: I1127 17:10:57.970976 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:57Z","lastTransitionTime":"2025-11-27T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.073941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.074078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.074100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.074164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.074184 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:58Z","lastTransitionTime":"2025-11-27T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.176865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.176912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.176930 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.176954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.176973 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:58Z","lastTransitionTime":"2025-11-27T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.279467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.279507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.279519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.279538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.279550 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:58Z","lastTransitionTime":"2025-11-27T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.381990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.382055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.382075 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.382101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.382118 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:58Z","lastTransitionTime":"2025-11-27T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.457418 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:10:58 crc kubenswrapper[4809]: E1127 17:10:58.457601 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.484430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.484477 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.484495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.484522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.484544 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:58Z","lastTransitionTime":"2025-11-27T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.587309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.587621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.587697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.587800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.587871 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:58Z","lastTransitionTime":"2025-11-27T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.690830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.690876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.690888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.690908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.690923 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:58Z","lastTransitionTime":"2025-11-27T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.793671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.793791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.793808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.793831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.793846 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:58Z","lastTransitionTime":"2025-11-27T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.897842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.897899 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.897912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.897933 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:58 crc kubenswrapper[4809]: I1127 17:10:58.897945 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:58Z","lastTransitionTime":"2025-11-27T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.001282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.001335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.001349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.001370 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.001385 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.104541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.104603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.104621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.104645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.104661 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.208533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.208644 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.208671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.208707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.208732 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.311854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.311918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.311934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.311957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.311971 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.415380 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.415435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.415447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.415466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.415480 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.457937 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:10:59 crc kubenswrapper[4809]: E1127 17:10:59.458109 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.458200 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.458304 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:10:59 crc kubenswrapper[4809]: E1127 17:10:59.458631 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:10:59 crc kubenswrapper[4809]: E1127 17:10:59.458712 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.518857 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.519004 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.519018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.519039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.519055 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.622803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.622874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.622894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.622925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.622947 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.726235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.726323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.726341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.726372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.726394 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.829235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.829281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.829292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.829306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.829317 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.932371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.932480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.932510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.932553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:10:59 crc kubenswrapper[4809]: I1127 17:10:59.932575 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:10:59Z","lastTransitionTime":"2025-11-27T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.036057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.036123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.036140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.036167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.036187 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.138487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.138533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.138543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.138562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.138576 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.241733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.241815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.241825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.241844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.241855 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.345301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.345372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.345395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.345431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.345456 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.449344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.449401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.449415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.449438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.449453 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.457626 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:00 crc kubenswrapper[4809]: E1127 17:11:00.457752 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.552507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.552556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.552568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.552585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.552597 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.656078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.656135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.656144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.656163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.656175 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.759236 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.759296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.759309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.759329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.759341 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.861835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.861910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.861920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.861937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.861947 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.964449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.964480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.964491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.964507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:00 crc kubenswrapper[4809]: I1127 17:11:00.964518 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:00Z","lastTransitionTime":"2025-11-27T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.066922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.066970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.066982 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.067003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.067015 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.169198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.169269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.169280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.169297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.169309 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.271855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.271914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.271936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.271964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.271984 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.375094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.375154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.375171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.375194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.375209 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.457896 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.457911 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.458135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:01 crc kubenswrapper[4809]: E1127 17:11:01.458487 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:01 crc kubenswrapper[4809]: E1127 17:11:01.458601 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:01 crc kubenswrapper[4809]: E1127 17:11:01.458784 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.477458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.477501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.477517 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.477535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.477549 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.579681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.579719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.579728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.579757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.579769 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.682226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.682263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.682276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.682292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.682302 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.784680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.784715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.784723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.784754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.784768 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.887392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.887451 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.887462 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.887487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.887498 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.990095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.990138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.990147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.990165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:01 crc kubenswrapper[4809]: I1127 17:11:01.990174 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:01Z","lastTransitionTime":"2025-11-27T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.092311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.092346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.092355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.092371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.092381 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:02Z","lastTransitionTime":"2025-11-27T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.194929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.194972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.194982 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.194998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.195008 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:02Z","lastTransitionTime":"2025-11-27T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.297986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.298060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.298072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.298094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.298107 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:02Z","lastTransitionTime":"2025-11-27T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.400895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.400974 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.400992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.401018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.401036 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:02Z","lastTransitionTime":"2025-11-27T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.457682 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:02 crc kubenswrapper[4809]: E1127 17:11:02.458263 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.458566 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:11:02 crc kubenswrapper[4809]: E1127 17:11:02.458783 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.504214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.504273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.504285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.504307 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.504319 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:02Z","lastTransitionTime":"2025-11-27T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.607266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.607345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.607368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.607392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.607408 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:02Z","lastTransitionTime":"2025-11-27T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.709980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.710116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.710140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.710174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.710198 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:02Z","lastTransitionTime":"2025-11-27T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.812960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.813028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.813040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.813072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.813098 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:02Z","lastTransitionTime":"2025-11-27T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.874193 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:02 crc kubenswrapper[4809]: E1127 17:11:02.874469 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:11:02 crc kubenswrapper[4809]: E1127 17:11:02.874603 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs podName:683f4d0e-67aa-40a7-b7b0-e5325ed0225f nodeName:}" failed. No retries permitted until 2025-11-27 17:12:06.8745698 +0000 UTC m=+162.147027162 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs") pod "network-metrics-daemon-2nc2d" (UID: "683f4d0e-67aa-40a7-b7b0-e5325ed0225f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.916069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.916118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.916132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.916155 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:02 crc kubenswrapper[4809]: I1127 17:11:02.916175 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:02Z","lastTransitionTime":"2025-11-27T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.017927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.018001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.018020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.018043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.018061 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.121401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.121469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.121485 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.121510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.121527 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.224848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.224922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.224940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.224966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.224982 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.328505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.328578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.328595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.328613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.328625 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.432232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.432289 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.432299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.432321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.432333 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.457154 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.457230 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.457168 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:03 crc kubenswrapper[4809]: E1127 17:11:03.457376 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:03 crc kubenswrapper[4809]: E1127 17:11:03.457478 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:03 crc kubenswrapper[4809]: E1127 17:11:03.457597 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.535961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.536043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.536058 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.536082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.536099 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.639190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.639265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.639285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.639332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.639351 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.743056 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.743114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.743130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.743155 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.743172 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.846793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.846841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.846853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.846872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.846885 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.950221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.950295 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.950320 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.950353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:03 crc kubenswrapper[4809]: I1127 17:11:03.950378 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:03Z","lastTransitionTime":"2025-11-27T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.053849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.054317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.054421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.054528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.054632 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.157707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.157806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.157818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.157837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.157852 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.261226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.261282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.261299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.261324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.261344 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.365278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.365322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.365335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.365355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.365369 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.457140 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:04 crc kubenswrapper[4809]: E1127 17:11:04.457382 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.468972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.469041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.469066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.469097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.469123 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.572733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.572795 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.572806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.572819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.572833 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.676000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.676076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.676089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.676114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.676127 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.779109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.779168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.779178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.779197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.779209 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.882214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.882268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.882277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.882298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.882309 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.984973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.985030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.985048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.985069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:04 crc kubenswrapper[4809]: I1127 17:11:04.985083 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:04Z","lastTransitionTime":"2025-11-27T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.087459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.087547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.087579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.087614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.087640 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:05Z","lastTransitionTime":"2025-11-27T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.190835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.190880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.190891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.190910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.190921 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:05Z","lastTransitionTime":"2025-11-27T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.293108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.293190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.293203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.293226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.293244 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:05Z","lastTransitionTime":"2025-11-27T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.396489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.396553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.396571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.396597 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.396616 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:05Z","lastTransitionTime":"2025-11-27T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.457342 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.457409 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:05 crc kubenswrapper[4809]: E1127 17:11:05.457601 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.457620 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:05 crc kubenswrapper[4809]: E1127 17:11:05.457788 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:05 crc kubenswrapper[4809]: E1127 17:11:05.457879 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.499882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.499928 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.499937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.499954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.499965 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:05Z","lastTransitionTime":"2025-11-27T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.503097 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.503066958 podStartE2EDuration="45.503066958s" podCreationTimestamp="2025-11-27 17:10:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.502850642 +0000 UTC m=+100.775308004" watchObservedRunningTime="2025-11-27 17:11:05.503066958 +0000 UTC m=+100.775524350" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.543788 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-d6cbb" podStartSLOduration=81.543714301 podStartE2EDuration="1m21.543714301s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.543357302 +0000 UTC m=+100.815814654" watchObservedRunningTime="2025-11-27 17:11:05.543714301 +0000 UTC m=+100.816171663" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.602734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.602840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.602864 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.602894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.602916 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:05Z","lastTransitionTime":"2025-11-27T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.631424 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pt22c" podStartSLOduration=80.631393936 podStartE2EDuration="1m20.631393936s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.596041873 +0000 UTC m=+100.868499255" watchObservedRunningTime="2025-11-27 17:11:05.631393936 +0000 UTC m=+100.903851308" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.632099 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=11.632091314 podStartE2EDuration="11.632091314s" podCreationTimestamp="2025-11-27 17:10:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.6319356 +0000 UTC m=+100.904392952" watchObservedRunningTime="2025-11-27 17:11:05.632091314 +0000 UTC m=+100.904548686" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.669474 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=80.669455101 podStartE2EDuration="1m20.669455101s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.65465739 +0000 UTC m=+100.927114742" watchObservedRunningTime="2025-11-27 17:11:05.669455101 +0000 UTC m=+100.941912453" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.693339 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=16.693311221 podStartE2EDuration="16.693311221s" podCreationTimestamp="2025-11-27 17:10:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.669365549 +0000 UTC m=+100.941822901" watchObservedRunningTime="2025-11-27 17:11:05.693311221 +0000 UTC m=+100.965768583" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.706071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.706125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.706136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.706161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.706174 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:05Z","lastTransitionTime":"2025-11-27T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.728630 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podStartSLOduration=81.728603753 podStartE2EDuration="1m21.728603753s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.708944744 +0000 UTC m=+100.981402096" watchObservedRunningTime="2025-11-27 17:11:05.728603753 +0000 UTC m=+101.001061115" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.729043 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=77.729037304 podStartE2EDuration="1m17.729037304s" podCreationTimestamp="2025-11-27 17:09:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.728528991 +0000 UTC m=+101.000986343" watchObservedRunningTime="2025-11-27 17:11:05.729037304 +0000 UTC m=+101.001494676" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.743565 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-fr7qb" podStartSLOduration=81.743533367 podStartE2EDuration="1m21.743533367s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.742376026 +0000 UTC m=+101.014833378" watchObservedRunningTime="2025-11-27 17:11:05.743533367 +0000 UTC m=+101.015990729" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.802798 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-s4g6j" podStartSLOduration=81.80274185 podStartE2EDuration="1m21.80274185s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.789920522 +0000 UTC m=+101.062377894" watchObservedRunningTime="2025-11-27 17:11:05.80274185 +0000 UTC m=+101.075199232" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.808763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.808824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.808837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.808859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.808873 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:05Z","lastTransitionTime":"2025-11-27T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.911073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.911111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.911120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.911135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:05 crc kubenswrapper[4809]: I1127 17:11:05.911145 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:05Z","lastTransitionTime":"2025-11-27T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.014403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.014472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.014489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.014509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.014519 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.118085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.118147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.118158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.118173 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.118182 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.221065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.221150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.221169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.221202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.221230 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.324647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.324719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.324733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.324786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.324807 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.428264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.428328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.428343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.428373 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.428391 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.457643 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:06 crc kubenswrapper[4809]: E1127 17:11:06.457939 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.531538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.531591 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.531604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.531622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.531633 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.635015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.635113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.635129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.635154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.635173 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.738817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.738869 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.738886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.738908 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.738924 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.842528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.842584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.842596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.842616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.842630 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.945557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.945626 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.945641 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.945667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:06 crc kubenswrapper[4809]: I1127 17:11:06.945683 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:06Z","lastTransitionTime":"2025-11-27T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.049474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.049550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.049567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.049595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.049611 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:07Z","lastTransitionTime":"2025-11-27T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.051409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.051443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.051454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.051474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.051487 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T17:11:07Z","lastTransitionTime":"2025-11-27T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.117333 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-gd8px" podStartSLOduration=83.117304777 podStartE2EDuration="1m23.117304777s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:05.832993449 +0000 UTC m=+101.105450801" watchObservedRunningTime="2025-11-27 17:11:07.117304777 +0000 UTC m=+102.389762129" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.118259 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7"] Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.118829 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.121358 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.122076 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.122340 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.122896 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.219695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0365759a-0383-4c60-b26e-9dd3ddfa448e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.220291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0365759a-0383-4c60-b26e-9dd3ddfa448e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.220445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0365759a-0383-4c60-b26e-9dd3ddfa448e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.220588 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0365759a-0383-4c60-b26e-9dd3ddfa448e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.220698 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0365759a-0383-4c60-b26e-9dd3ddfa448e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.321379 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0365759a-0383-4c60-b26e-9dd3ddfa448e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.321472 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0365759a-0383-4c60-b26e-9dd3ddfa448e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.321513 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0365759a-0383-4c60-b26e-9dd3ddfa448e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.321550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0365759a-0383-4c60-b26e-9dd3ddfa448e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.321643 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0365759a-0383-4c60-b26e-9dd3ddfa448e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.322269 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0365759a-0383-4c60-b26e-9dd3ddfa448e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.322522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0365759a-0383-4c60-b26e-9dd3ddfa448e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.323440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0365759a-0383-4c60-b26e-9dd3ddfa448e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.331335 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0365759a-0383-4c60-b26e-9dd3ddfa448e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.350619 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0365759a-0383-4c60-b26e-9dd3ddfa448e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-t78z7\" (UID: \"0365759a-0383-4c60-b26e-9dd3ddfa448e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.443803 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.457563 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.457716 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:07 crc kubenswrapper[4809]: E1127 17:11:07.457881 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:07 crc kubenswrapper[4809]: I1127 17:11:07.457637 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:07 crc kubenswrapper[4809]: E1127 17:11:07.458102 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:07 crc kubenswrapper[4809]: E1127 17:11:07.457995 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:07 crc kubenswrapper[4809]: W1127 17:11:07.469772 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0365759a_0383_4c60_b26e_9dd3ddfa448e.slice/crio-e432fb53b042a0718716c9c30f351c8518613bed53f29073772fe46943132124 WatchSource:0}: Error finding container e432fb53b042a0718716c9c30f351c8518613bed53f29073772fe46943132124: Status 404 returned error can't find the container with id e432fb53b042a0718716c9c30f351c8518613bed53f29073772fe46943132124 Nov 27 17:11:08 crc kubenswrapper[4809]: I1127 17:11:08.037990 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" event={"ID":"0365759a-0383-4c60-b26e-9dd3ddfa448e","Type":"ContainerStarted","Data":"00a9549be2f84664bb155d3652d714597a221344b8c079e5da3b409e3e7b8178"} Nov 27 17:11:08 crc kubenswrapper[4809]: I1127 17:11:08.038063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" event={"ID":"0365759a-0383-4c60-b26e-9dd3ddfa448e","Type":"ContainerStarted","Data":"e432fb53b042a0718716c9c30f351c8518613bed53f29073772fe46943132124"} Nov 27 17:11:08 crc kubenswrapper[4809]: I1127 17:11:08.062595 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t78z7" podStartSLOduration=84.062572153 podStartE2EDuration="1m24.062572153s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:08.061608438 +0000 UTC m=+103.334065830" watchObservedRunningTime="2025-11-27 17:11:08.062572153 +0000 UTC m=+103.335029505" Nov 27 17:11:08 crc kubenswrapper[4809]: I1127 17:11:08.457724 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:08 crc kubenswrapper[4809]: E1127 17:11:08.457918 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:09 crc kubenswrapper[4809]: I1127 17:11:09.457078 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:09 crc kubenswrapper[4809]: E1127 17:11:09.457215 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:09 crc kubenswrapper[4809]: I1127 17:11:09.457107 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:09 crc kubenswrapper[4809]: I1127 17:11:09.457079 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:09 crc kubenswrapper[4809]: E1127 17:11:09.457344 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:09 crc kubenswrapper[4809]: E1127 17:11:09.457441 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:10 crc kubenswrapper[4809]: I1127 17:11:10.457340 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:10 crc kubenswrapper[4809]: E1127 17:11:10.457961 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:11 crc kubenswrapper[4809]: I1127 17:11:11.457280 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:11 crc kubenswrapper[4809]: I1127 17:11:11.457368 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:11 crc kubenswrapper[4809]: I1127 17:11:11.457427 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:11 crc kubenswrapper[4809]: E1127 17:11:11.457559 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:11 crc kubenswrapper[4809]: E1127 17:11:11.457660 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:11 crc kubenswrapper[4809]: E1127 17:11:11.457822 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:12 crc kubenswrapper[4809]: I1127 17:11:12.457884 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:12 crc kubenswrapper[4809]: E1127 17:11:12.458115 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:13 crc kubenswrapper[4809]: I1127 17:11:13.457266 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:13 crc kubenswrapper[4809]: E1127 17:11:13.457508 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:13 crc kubenswrapper[4809]: I1127 17:11:13.457619 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:13 crc kubenswrapper[4809]: I1127 17:11:13.457670 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:13 crc kubenswrapper[4809]: E1127 17:11:13.457920 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:13 crc kubenswrapper[4809]: E1127 17:11:13.458006 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:13 crc kubenswrapper[4809]: I1127 17:11:13.458271 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:11:13 crc kubenswrapper[4809]: E1127 17:11:13.458419 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-m5b98_openshift-ovn-kubernetes(66b4318a-f089-451d-8a16-97de26acce28)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" Nov 27 17:11:14 crc kubenswrapper[4809]: I1127 17:11:14.457164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:14 crc kubenswrapper[4809]: E1127 17:11:14.457315 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:15 crc kubenswrapper[4809]: I1127 17:11:15.456876 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:15 crc kubenswrapper[4809]: I1127 17:11:15.456935 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:15 crc kubenswrapper[4809]: I1127 17:11:15.457005 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:15 crc kubenswrapper[4809]: E1127 17:11:15.458073 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:15 crc kubenswrapper[4809]: E1127 17:11:15.458252 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:15 crc kubenswrapper[4809]: E1127 17:11:15.458353 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:16 crc kubenswrapper[4809]: I1127 17:11:16.457295 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:16 crc kubenswrapper[4809]: E1127 17:11:16.457659 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:17 crc kubenswrapper[4809]: I1127 17:11:17.456977 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:17 crc kubenswrapper[4809]: I1127 17:11:17.457031 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:17 crc kubenswrapper[4809]: I1127 17:11:17.457093 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:17 crc kubenswrapper[4809]: E1127 17:11:17.457147 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:17 crc kubenswrapper[4809]: E1127 17:11:17.457265 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:17 crc kubenswrapper[4809]: E1127 17:11:17.457363 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:18 crc kubenswrapper[4809]: I1127 17:11:18.457184 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:18 crc kubenswrapper[4809]: E1127 17:11:18.457415 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:19 crc kubenswrapper[4809]: I1127 17:11:19.457361 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:19 crc kubenswrapper[4809]: I1127 17:11:19.457493 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:19 crc kubenswrapper[4809]: I1127 17:11:19.457362 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:19 crc kubenswrapper[4809]: E1127 17:11:19.457549 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:19 crc kubenswrapper[4809]: E1127 17:11:19.457641 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:19 crc kubenswrapper[4809]: E1127 17:11:19.457791 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:20 crc kubenswrapper[4809]: I1127 17:11:20.457417 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:20 crc kubenswrapper[4809]: E1127 17:11:20.457638 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:21 crc kubenswrapper[4809]: I1127 17:11:21.080823 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd8px_8741a6e8-a7f2-40ef-b25b-10e518345478/kube-multus/1.log" Nov 27 17:11:21 crc kubenswrapper[4809]: I1127 17:11:21.081590 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd8px_8741a6e8-a7f2-40ef-b25b-10e518345478/kube-multus/0.log" Nov 27 17:11:21 crc kubenswrapper[4809]: I1127 17:11:21.081631 4809 generic.go:334] "Generic (PLEG): container finished" podID="8741a6e8-a7f2-40ef-b25b-10e518345478" containerID="aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156" exitCode=1 Nov 27 17:11:21 crc kubenswrapper[4809]: I1127 17:11:21.081663 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd8px" event={"ID":"8741a6e8-a7f2-40ef-b25b-10e518345478","Type":"ContainerDied","Data":"aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156"} Nov 27 17:11:21 crc kubenswrapper[4809]: I1127 17:11:21.081704 4809 scope.go:117] "RemoveContainer" containerID="e9999bd2e297bb8b867bbcf12767fcb6e7885feee518577d96a831b86de01440" Nov 27 17:11:21 crc kubenswrapper[4809]: I1127 17:11:21.082573 4809 scope.go:117] "RemoveContainer" containerID="aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156" Nov 27 17:11:21 crc kubenswrapper[4809]: E1127 17:11:21.083612 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-gd8px_openshift-multus(8741a6e8-a7f2-40ef-b25b-10e518345478)\"" pod="openshift-multus/multus-gd8px" podUID="8741a6e8-a7f2-40ef-b25b-10e518345478" Nov 27 17:11:21 crc kubenswrapper[4809]: I1127 17:11:21.456988 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:21 crc kubenswrapper[4809]: I1127 17:11:21.457066 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:21 crc kubenswrapper[4809]: I1127 17:11:21.457102 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:21 crc kubenswrapper[4809]: E1127 17:11:21.457783 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:21 crc kubenswrapper[4809]: E1127 17:11:21.457906 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:21 crc kubenswrapper[4809]: E1127 17:11:21.458098 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:22 crc kubenswrapper[4809]: I1127 17:11:22.087926 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd8px_8741a6e8-a7f2-40ef-b25b-10e518345478/kube-multus/1.log" Nov 27 17:11:22 crc kubenswrapper[4809]: I1127 17:11:22.457201 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:22 crc kubenswrapper[4809]: E1127 17:11:22.457373 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:23 crc kubenswrapper[4809]: I1127 17:11:23.457389 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:23 crc kubenswrapper[4809]: I1127 17:11:23.457439 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:23 crc kubenswrapper[4809]: E1127 17:11:23.457541 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:23 crc kubenswrapper[4809]: I1127 17:11:23.457561 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:23 crc kubenswrapper[4809]: E1127 17:11:23.457722 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:23 crc kubenswrapper[4809]: E1127 17:11:23.458008 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:24 crc kubenswrapper[4809]: I1127 17:11:24.456951 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:24 crc kubenswrapper[4809]: E1127 17:11:24.457175 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:25 crc kubenswrapper[4809]: E1127 17:11:25.394065 4809 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 27 17:11:25 crc kubenswrapper[4809]: I1127 17:11:25.457064 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:25 crc kubenswrapper[4809]: I1127 17:11:25.457177 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:25 crc kubenswrapper[4809]: E1127 17:11:25.458133 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:25 crc kubenswrapper[4809]: I1127 17:11:25.458351 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:25 crc kubenswrapper[4809]: E1127 17:11:25.458482 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:25 crc kubenswrapper[4809]: E1127 17:11:25.458781 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:25 crc kubenswrapper[4809]: E1127 17:11:25.562500 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 17:11:26 crc kubenswrapper[4809]: I1127 17:11:26.457480 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:26 crc kubenswrapper[4809]: E1127 17:11:26.457651 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:27 crc kubenswrapper[4809]: I1127 17:11:27.457155 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:27 crc kubenswrapper[4809]: I1127 17:11:27.457210 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:27 crc kubenswrapper[4809]: E1127 17:11:27.457285 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:27 crc kubenswrapper[4809]: I1127 17:11:27.457392 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:27 crc kubenswrapper[4809]: I1127 17:11:27.458088 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:11:27 crc kubenswrapper[4809]: E1127 17:11:27.458201 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:27 crc kubenswrapper[4809]: E1127 17:11:27.458328 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:28 crc kubenswrapper[4809]: I1127 17:11:28.111004 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/3.log" Nov 27 17:11:28 crc kubenswrapper[4809]: I1127 17:11:28.113550 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerStarted","Data":"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b"} Nov 27 17:11:28 crc kubenswrapper[4809]: I1127 17:11:28.114068 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:11:28 crc kubenswrapper[4809]: I1127 17:11:28.142967 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podStartSLOduration=103.142948464 podStartE2EDuration="1m43.142948464s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:28.141780073 +0000 UTC m=+123.414237435" watchObservedRunningTime="2025-11-27 17:11:28.142948464 +0000 UTC m=+123.415405816" Nov 27 17:11:28 crc kubenswrapper[4809]: I1127 17:11:28.457417 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:28 crc kubenswrapper[4809]: E1127 17:11:28.457581 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:28 crc kubenswrapper[4809]: I1127 17:11:28.476264 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2nc2d"] Nov 27 17:11:29 crc kubenswrapper[4809]: I1127 17:11:29.119299 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:29 crc kubenswrapper[4809]: E1127 17:11:29.120126 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:29 crc kubenswrapper[4809]: I1127 17:11:29.457526 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:29 crc kubenswrapper[4809]: I1127 17:11:29.457652 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:29 crc kubenswrapper[4809]: I1127 17:11:29.457688 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:29 crc kubenswrapper[4809]: E1127 17:11:29.457847 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:29 crc kubenswrapper[4809]: E1127 17:11:29.458231 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:29 crc kubenswrapper[4809]: E1127 17:11:29.458550 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:30 crc kubenswrapper[4809]: E1127 17:11:30.564101 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 17:11:31 crc kubenswrapper[4809]: I1127 17:11:31.457458 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:31 crc kubenswrapper[4809]: I1127 17:11:31.457518 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:31 crc kubenswrapper[4809]: I1127 17:11:31.457560 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:31 crc kubenswrapper[4809]: I1127 17:11:31.457649 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:31 crc kubenswrapper[4809]: E1127 17:11:31.457666 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:31 crc kubenswrapper[4809]: E1127 17:11:31.457836 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:31 crc kubenswrapper[4809]: E1127 17:11:31.457945 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:31 crc kubenswrapper[4809]: E1127 17:11:31.458040 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:32 crc kubenswrapper[4809]: I1127 17:11:32.457427 4809 scope.go:117] "RemoveContainer" containerID="aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156" Nov 27 17:11:33 crc kubenswrapper[4809]: I1127 17:11:33.135105 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd8px_8741a6e8-a7f2-40ef-b25b-10e518345478/kube-multus/1.log" Nov 27 17:11:33 crc kubenswrapper[4809]: I1127 17:11:33.135569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd8px" event={"ID":"8741a6e8-a7f2-40ef-b25b-10e518345478","Type":"ContainerStarted","Data":"090893af68773017e5a04393586f2d8375589b6156f7fca536afd90c178c9f26"} Nov 27 17:11:33 crc kubenswrapper[4809]: I1127 17:11:33.457438 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:33 crc kubenswrapper[4809]: I1127 17:11:33.457472 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:33 crc kubenswrapper[4809]: E1127 17:11:33.457617 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:33 crc kubenswrapper[4809]: I1127 17:11:33.457702 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:33 crc kubenswrapper[4809]: I1127 17:11:33.457702 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:33 crc kubenswrapper[4809]: E1127 17:11:33.457909 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:33 crc kubenswrapper[4809]: E1127 17:11:33.458041 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:33 crc kubenswrapper[4809]: E1127 17:11:33.458194 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:35 crc kubenswrapper[4809]: I1127 17:11:35.457494 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:35 crc kubenswrapper[4809]: I1127 17:11:35.457963 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:35 crc kubenswrapper[4809]: E1127 17:11:35.457938 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 17:11:35 crc kubenswrapper[4809]: I1127 17:11:35.457984 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:35 crc kubenswrapper[4809]: I1127 17:11:35.458025 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:35 crc kubenswrapper[4809]: E1127 17:11:35.460386 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 17:11:35 crc kubenswrapper[4809]: E1127 17:11:35.460464 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 17:11:35 crc kubenswrapper[4809]: E1127 17:11:35.460584 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nc2d" podUID="683f4d0e-67aa-40a7-b7b0-e5325ed0225f" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.457798 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.457830 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.457849 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.458006 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.463544 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.463883 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.464200 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.464262 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.464277 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.464218 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.630657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.680343 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.681303 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.684542 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.685108 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.685591 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.686710 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.687168 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.687340 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.688651 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6pm2s"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.689265 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.689415 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.689528 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.690666 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.692263 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.692528 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2cphh"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.692701 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.692907 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.695154 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.695327 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hgmrs"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.696001 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.699308 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.699433 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.701279 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.701486 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.704952 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.707497 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.708150 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.708638 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.708827 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.708988 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.709080 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.709198 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.709403 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6rxcc"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.709527 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.709080 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.709922 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.710014 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.710104 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.710203 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.710345 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.710952 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.732277 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.740590 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.741943 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.742212 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.758656 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-w7f6q"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.759282 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.759636 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.759813 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.760190 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.760303 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.759279 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.760724 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.761010 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.761187 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.761529 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7c7gv"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.759613 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.759613 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.760092 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.761995 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.762775 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.763880 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2t2fb"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.764215 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gzfwc"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.764335 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.764338 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.764711 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.765050 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.765143 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-jtxfb"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.765537 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jtxfb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.765568 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.765648 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.765695 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.765861 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767691 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54856698-dafa-4315-84c3-e0b746e815f9-audit-policies\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767765 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767811 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb38e0de-3758-4e46-beb3-cf27c411725e-service-ca-bundle\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767832 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-config\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-etcd-serving-ca\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767872 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-node-pullsecrets\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-image-import-ca\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767912 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-config\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767931 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-audit-dir\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.767963 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768040 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-serving-cert\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768280 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54856698-dafa-4315-84c3-e0b746e815f9-audit-dir\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768361 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/54856698-dafa-4315-84c3-e0b746e815f9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b5rd\" (UniqueName: \"kubernetes.io/projected/23af03b4-a427-4e08-848b-0561b9cc7364-kube-api-access-9b5rd\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768460 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb38e0de-3758-4e46-beb3-cf27c411725e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768514 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-audit\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768540 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-etcd-client\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768620 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxfwr\" (UniqueName: \"kubernetes.io/projected/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-kube-api-access-cxfwr\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768684 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54856698-dafa-4315-84c3-e0b746e815f9-serving-cert\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768785 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-client-ca\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768822 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23af03b4-a427-4e08-848b-0561b9cc7364-serving-cert\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768877 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb38e0de-3758-4e46-beb3-cf27c411725e-config\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768936 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmbqq\" (UniqueName: \"kubernetes.io/projected/54856698-dafa-4315-84c3-e0b746e815f9-kube-api-access-vmbqq\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.768971 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54856698-dafa-4315-84c3-e0b746e815f9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.769012 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/54856698-dafa-4315-84c3-e0b746e815f9-encryption-config\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.769061 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54856698-dafa-4315-84c3-e0b746e815f9-etcd-client\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.769107 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb38e0de-3758-4e46-beb3-cf27c411725e-serving-cert\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.769160 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54jb7\" (UniqueName: \"kubernetes.io/projected/cb38e0de-3758-4e46-beb3-cf27c411725e-kube-api-access-54jb7\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.769186 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-encryption-config\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.769614 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.770108 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.774570 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.775209 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.775591 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.775931 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.777066 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-m5lpz"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.777880 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dj24l"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.778294 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.778337 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.778518 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.778632 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.778733 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.778795 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.780174 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.780842 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.781277 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.781434 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.781851 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.782004 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.784102 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.784239 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.784891 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785230 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785393 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785447 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785490 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785602 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785631 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785652 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785721 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785789 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785860 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.785880 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.786476 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.786569 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.786695 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.786863 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.786945 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.787063 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.787252 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.787312 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.787351 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.787677 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.790506 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.792765 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.793501 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.794000 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.794181 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.794323 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.794423 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.794498 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.794684 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.794767 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.794835 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.794965 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.804102 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.804399 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.804491 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.804532 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.804871 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.805245 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.805267 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.805867 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.806475 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.807397 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.817790 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.825820 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.825939 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.834006 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.835479 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.835619 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.836036 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.843791 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.844369 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.844488 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.844522 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.845015 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.845050 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.845536 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mcsp5"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.845996 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.846118 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.847013 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.847966 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.849161 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.849229 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.849488 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.851396 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.851916 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.852573 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.853141 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.853409 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.853545 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.855181 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.855564 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.856241 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.856940 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.857914 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.858771 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.860004 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-kzdxz"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.860698 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.862407 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-94nbr"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.863218 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.863524 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8lfbq"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.864063 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.865365 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6pm2s"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.866264 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.866848 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.867660 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.868033 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.868636 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.869207 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-r7lqs"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.869880 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-client-ca\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.869995 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-r7lqs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.869999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r46j\" (UniqueName: \"kubernetes.io/projected/fff4c3fa-a124-4cb5-806d-eea27e5dfd4a-kube-api-access-4r46j\") pod \"downloads-7954f5f757-jtxfb\" (UID: \"fff4c3fa-a124-4cb5-806d-eea27e5dfd4a\") " pod="openshift-console/downloads-7954f5f757-jtxfb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.870188 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.870286 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jwdb\" (UniqueName: \"kubernetes.io/projected/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-kube-api-access-2jwdb\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.870730 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d0a292-6e77-4197-9877-72e8fae6b941-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xbsxf\" (UID: \"08d0a292-6e77-4197-9877-72e8fae6b941\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.870863 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9lp5\" (UniqueName: \"kubernetes.io/projected/d9a53c5a-f3a7-484e-9058-11e61da9c9ed-kube-api-access-l9lp5\") pod \"openshift-config-operator-7777fb866f-mqxzb\" (UID: \"d9a53c5a-f3a7-484e-9058-11e61da9c9ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.870984 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23af03b4-a427-4e08-848b-0561b9cc7364-serving-cert\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb38e0de-3758-4e46-beb3-cf27c411725e-config\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871189 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dccff42e-1a6b-4239-a095-35af760b79b1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7mdsb\" (UID: \"dccff42e-1a6b-4239-a095-35af760b79b1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871281 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9a53c5a-f3a7-484e-9058-11e61da9c9ed-serving-cert\") pod \"openshift-config-operator-7777fb866f-mqxzb\" (UID: \"d9a53c5a-f3a7-484e-9058-11e61da9c9ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871376 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-machine-approver-tls\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871482 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmbqq\" (UniqueName: \"kubernetes.io/projected/54856698-dafa-4315-84c3-e0b746e815f9-kube-api-access-vmbqq\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871584 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkr26\" (UniqueName: \"kubernetes.io/projected/02f0a77f-7099-480e-b4e6-6183592ca5e6-kube-api-access-hkr26\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871716 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d397f9a-119c-4d88-ae8d-68ed99c919f3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871718 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6v9pw"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871897 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54856698-dafa-4315-84c3-e0b746e815f9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871940 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/54856698-dafa-4315-84c3-e0b746e815f9-encryption-config\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.871972 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzf5d\" (UniqueName: \"kubernetes.io/projected/93429dc0-c4aa-4eec-a48d-7128e13d0d8f-kube-api-access-dzf5d\") pod \"openshift-controller-manager-operator-756b6f6bc6-kwxf5\" (UID: \"93429dc0-c4aa-4eec-a48d-7128e13d0d8f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872001 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dccff42e-1a6b-4239-a095-35af760b79b1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7mdsb\" (UID: \"dccff42e-1a6b-4239-a095-35af760b79b1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872029 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60414207-aa6c-42ec-a02c-562a10239881-config\") pod \"kube-controller-manager-operator-78b949d7b-kzwwp\" (UID: \"60414207-aa6c-42ec-a02c-562a10239881\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872050 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/86022daa-0ebf-4820-a476-e5cacc43f9d5-metrics-tls\") pod \"dns-operator-744455d44c-gzfwc\" (UID: \"86022daa-0ebf-4820-a476-e5cacc43f9d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872070 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d94e93e8-457e-4b40-b09d-480ff253bb8a-trusted-ca\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872093 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-oauth-serving-cert\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872118 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94e93e8-457e-4b40-b09d-480ff253bb8a-config\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872140 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02f0a77f-7099-480e-b4e6-6183592ca5e6-config\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872160 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9vkq\" (UniqueName: \"kubernetes.io/projected/68ba89bb-2ac5-4204-8362-891783709611-kube-api-access-x9vkq\") pod \"openshift-apiserver-operator-796bbdcf4f-hxvj9\" (UID: \"68ba89bb-2ac5-4204-8362-891783709611\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54856698-dafa-4315-84c3-e0b746e815f9-etcd-client\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872208 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872226 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872247 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbtm7\" (UniqueName: \"kubernetes.io/projected/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-kube-api-access-sbtm7\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872264 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02f0a77f-7099-480e-b4e6-6183592ca5e6-serving-cert\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872283 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872303 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dptq\" (UniqueName: \"kubernetes.io/projected/08d0a292-6e77-4197-9877-72e8fae6b941-kube-api-access-2dptq\") pod \"kube-storage-version-migrator-operator-b67b599dd-xbsxf\" (UID: \"08d0a292-6e77-4197-9877-72e8fae6b941\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872325 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-metrics-tls\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872375 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60414207-aa6c-42ec-a02c-562a10239881-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kzwwp\" (UID: \"60414207-aa6c-42ec-a02c-562a10239881\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-oauth-config\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb38e0de-3758-4e46-beb3-cf27c411725e-config\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872480 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93429dc0-c4aa-4eec-a48d-7128e13d0d8f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kwxf5\" (UID: \"93429dc0-c4aa-4eec-a48d-7128e13d0d8f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872514 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872547 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb38e0de-3758-4e46-beb3-cf27c411725e-serving-cert\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872569 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-client-ca\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872599 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54jb7\" (UniqueName: \"kubernetes.io/projected/cb38e0de-3758-4e46-beb3-cf27c411725e-kube-api-access-54jb7\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872625 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-encryption-config\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872871 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dccff42e-1a6b-4239-a095-35af760b79b1-config\") pod \"kube-apiserver-operator-766d6c64bb-7mdsb\" (UID: \"dccff42e-1a6b-4239-a095-35af760b79b1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.872972 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93429dc0-c4aa-4eec-a48d-7128e13d0d8f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kwxf5\" (UID: \"93429dc0-c4aa-4eec-a48d-7128e13d0d8f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873011 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873037 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvtqt\" (UniqueName: \"kubernetes.io/projected/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-kube-api-access-kvtqt\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873057 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873080 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-client-ca\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54856698-dafa-4315-84c3-e0b746e815f9-audit-policies\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873121 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s4lpp\" (UID: \"d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/02f0a77f-7099-480e-b4e6-6183592ca5e6-etcd-client\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpsk8\" (UniqueName: \"kubernetes.io/projected/d94e93e8-457e-4b40-b09d-480ff253bb8a-kube-api-access-mpsk8\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873193 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb38e0de-3758-4e46-beb3-cf27c411725e-service-ca-bundle\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873212 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-config\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873232 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-serving-cert\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873252 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873272 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873290 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2d397f9a-119c-4d88-ae8d-68ed99c919f3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-auth-proxy-config\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873327 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-config\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873346 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873366 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-etcd-serving-ca\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873381 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbb4h\" (UniqueName: \"kubernetes.io/projected/fb3ec517-319b-4bd9-87b9-19ee21f410b6-kube-api-access-qbb4h\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-image-import-ca\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873420 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-images\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873437 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873455 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ba89bb-2ac5-4204-8362-891783709611-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hxvj9\" (UID: \"68ba89bb-2ac5-4204-8362-891783709611\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873473 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-node-pullsecrets\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873516 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-config\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873533 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-audit-dir\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-trusted-ca-bundle\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873571 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873595 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-serving-cert\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873612 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frzs8\" (UniqueName: \"kubernetes.io/projected/d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4-kube-api-access-frzs8\") pod \"cluster-samples-operator-665b6dd947-s4lpp\" (UID: \"d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873631 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-config\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873650 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg7hg\" (UniqueName: \"kubernetes.io/projected/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-kube-api-access-lg7hg\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873669 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfjj4\" (UniqueName: \"kubernetes.io/projected/2d397f9a-119c-4d88-ae8d-68ed99c919f3-kube-api-access-tfjj4\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873704 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08d0a292-6e77-4197-9877-72e8fae6b941-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xbsxf\" (UID: \"08d0a292-6e77-4197-9877-72e8fae6b941\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873726 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-dir\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873764 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873798 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54856698-dafa-4315-84c3-e0b746e815f9-audit-dir\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873818 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-trusted-ca\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873836 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/02f0a77f-7099-480e-b4e6-6183592ca5e6-etcd-service-ca\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873855 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-policies\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874190 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwvlz\" (UniqueName: \"kubernetes.io/projected/f6349461-b5ae-4e1a-9711-9d676f1ec01f-kube-api-access-pwvlz\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874214 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/54856698-dafa-4315-84c3-e0b746e815f9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874233 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b5rd\" (UniqueName: \"kubernetes.io/projected/23af03b4-a427-4e08-848b-0561b9cc7364-kube-api-access-9b5rd\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/02f0a77f-7099-480e-b4e6-6183592ca5e6-etcd-ca\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874265 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6349461-b5ae-4e1a-9711-9d676f1ec01f-serving-cert\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874280 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d9a53c5a-f3a7-484e-9058-11e61da9c9ed-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mqxzb\" (UID: \"d9a53c5a-f3a7-484e-9058-11e61da9c9ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874296 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb38e0de-3758-4e46-beb3-cf27c411725e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874311 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-service-ca\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874328 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60414207-aa6c-42ec-a02c-562a10239881-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kzwwp\" (UID: \"60414207-aa6c-42ec-a02c-562a10239881\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874348 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-audit\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874364 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-etcd-client\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874397 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxfwr\" (UniqueName: \"kubernetes.io/projected/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-kube-api-access-cxfwr\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874432 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-config\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874482 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb38e0de-3758-4e46-beb3-cf27c411725e-service-ca-bundle\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874506 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-node-pullsecrets\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874506 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-config\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874553 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d94e93e8-457e-4b40-b09d-480ff253bb8a-serving-cert\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874574 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d397f9a-119c-4d88-ae8d-68ed99c919f3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874602 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68ba89bb-2ac5-4204-8362-891783709611-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hxvj9\" (UID: \"68ba89bb-2ac5-4204-8362-891783709611\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874637 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54856698-dafa-4315-84c3-e0b746e815f9-serving-cert\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.874659 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llrhw\" (UniqueName: \"kubernetes.io/projected/86022daa-0ebf-4820-a476-e5cacc43f9d5-kube-api-access-llrhw\") pod \"dns-operator-744455d44c-gzfwc\" (UID: \"86022daa-0ebf-4820-a476-e5cacc43f9d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.875107 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-config\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.875406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-audit-dir\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.878955 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-etcd-serving-ca\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.879247 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54856698-dafa-4315-84c3-e0b746e815f9-audit-policies\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.873704 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54856698-dafa-4315-84c3-e0b746e815f9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.880270 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.880413 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23af03b4-a427-4e08-848b-0561b9cc7364-serving-cert\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.880674 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-encryption-config\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.880872 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/54856698-dafa-4315-84c3-e0b746e815f9-encryption-config\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.881233 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54856698-dafa-4315-84c3-e0b746e815f9-audit-dir\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.882303 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb38e0de-3758-4e46-beb3-cf27c411725e-serving-cert\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.882992 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6rxcc"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.883038 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.883175 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.883364 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-audit\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.884671 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-config\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.885803 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-image-import-ca\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.886061 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54856698-dafa-4315-84c3-e0b746e815f9-etcd-client\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.886767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/54856698-dafa-4315-84c3-e0b746e815f9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.887293 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.887461 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb38e0de-3758-4e46-beb3-cf27c411725e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.890761 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jtxfb"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.890836 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gzfwc"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.892681 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-etcd-client\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.893860 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-w7f6q"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.894458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54856698-dafa-4315-84c3-e0b746e815f9-serving-cert\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.894883 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-serving-cert\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.897370 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.898263 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.902054 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2t2fb"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.904471 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.904714 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.906069 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mcsp5"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.907959 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hgmrs"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.908489 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-m5lpz"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.909493 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.910533 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.911648 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.913063 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7c7gv"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.914133 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.915420 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2cphh"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.916428 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.917461 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.918869 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.920585 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dj24l"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.921622 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.922882 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.923091 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.924056 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.925065 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.926122 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.928234 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.929837 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.930349 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6v9pw"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.931532 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-94nbr"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.932659 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.934371 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.935771 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-zpwvm"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.936791 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.937524 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dpfxv"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.938572 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.938943 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.940470 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.941887 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8lfbq"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.943292 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-r7lqs"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.943679 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.944698 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dpfxv"] Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.964326 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976077 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976115 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dptq\" (UniqueName: \"kubernetes.io/projected/08d0a292-6e77-4197-9877-72e8fae6b941-kube-api-access-2dptq\") pod \"kube-storage-version-migrator-operator-b67b599dd-xbsxf\" (UID: \"08d0a292-6e77-4197-9877-72e8fae6b941\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976169 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbtm7\" (UniqueName: \"kubernetes.io/projected/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-kube-api-access-sbtm7\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02f0a77f-7099-480e-b4e6-6183592ca5e6-serving-cert\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976227 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-oauth-config\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976244 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-metrics-tls\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976260 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60414207-aa6c-42ec-a02c-562a10239881-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kzwwp\" (UID: \"60414207-aa6c-42ec-a02c-562a10239881\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93429dc0-c4aa-4eec-a48d-7128e13d0d8f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kwxf5\" (UID: \"93429dc0-c4aa-4eec-a48d-7128e13d0d8f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976296 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dccff42e-1a6b-4239-a095-35af760b79b1-config\") pod \"kube-apiserver-operator-766d6c64bb-7mdsb\" (UID: \"dccff42e-1a6b-4239-a095-35af760b79b1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976338 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93429dc0-c4aa-4eec-a48d-7128e13d0d8f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kwxf5\" (UID: \"93429dc0-c4aa-4eec-a48d-7128e13d0d8f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-client-ca\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976371 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvtqt\" (UniqueName: \"kubernetes.io/projected/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-kube-api-access-kvtqt\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976389 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s4lpp\" (UID: \"d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/02f0a77f-7099-480e-b4e6-6183592ca5e6-etcd-client\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976440 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpsk8\" (UniqueName: \"kubernetes.io/projected/d94e93e8-457e-4b40-b09d-480ff253bb8a-kube-api-access-mpsk8\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976466 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-serving-cert\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976483 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-auth-proxy-config\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976500 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-config\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976516 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976534 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976555 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2d397f9a-119c-4d88-ae8d-68ed99c919f3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976580 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976597 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbb4h\" (UniqueName: \"kubernetes.io/projected/fb3ec517-319b-4bd9-87b9-19ee21f410b6-kube-api-access-qbb4h\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976616 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ba89bb-2ac5-4204-8362-891783709611-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hxvj9\" (UID: \"68ba89bb-2ac5-4204-8362-891783709611\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976633 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-images\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976651 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976668 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-trusted-ca-bundle\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976685 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frzs8\" (UniqueName: \"kubernetes.io/projected/d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4-kube-api-access-frzs8\") pod \"cluster-samples-operator-665b6dd947-s4lpp\" (UID: \"d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976703 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-config\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg7hg\" (UniqueName: \"kubernetes.io/projected/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-kube-api-access-lg7hg\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976754 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfjj4\" (UniqueName: \"kubernetes.io/projected/2d397f9a-119c-4d88-ae8d-68ed99c919f3-kube-api-access-tfjj4\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976772 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08d0a292-6e77-4197-9877-72e8fae6b941-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xbsxf\" (UID: \"08d0a292-6e77-4197-9877-72e8fae6b941\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976789 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976808 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-dir\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976824 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976850 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/02f0a77f-7099-480e-b4e6-6183592ca5e6-etcd-service-ca\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976868 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-policies\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976909 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwvlz\" (UniqueName: \"kubernetes.io/projected/f6349461-b5ae-4e1a-9711-9d676f1ec01f-kube-api-access-pwvlz\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976929 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-trusted-ca\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976954 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/02f0a77f-7099-480e-b4e6-6183592ca5e6-etcd-ca\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6349461-b5ae-4e1a-9711-9d676f1ec01f-serving-cert\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.976989 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d9a53c5a-f3a7-484e-9058-11e61da9c9ed-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mqxzb\" (UID: \"d9a53c5a-f3a7-484e-9058-11e61da9c9ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977026 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-service-ca\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977045 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60414207-aa6c-42ec-a02c-562a10239881-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kzwwp\" (UID: \"60414207-aa6c-42ec-a02c-562a10239881\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977077 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977095 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-config\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977111 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-config\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977129 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d94e93e8-457e-4b40-b09d-480ff253bb8a-serving-cert\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977146 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d397f9a-119c-4d88-ae8d-68ed99c919f3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977166 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68ba89bb-2ac5-4204-8362-891783709611-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hxvj9\" (UID: \"68ba89bb-2ac5-4204-8362-891783709611\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llrhw\" (UniqueName: \"kubernetes.io/projected/86022daa-0ebf-4820-a476-e5cacc43f9d5-kube-api-access-llrhw\") pod \"dns-operator-744455d44c-gzfwc\" (UID: \"86022daa-0ebf-4820-a476-e5cacc43f9d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977220 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d0a292-6e77-4197-9877-72e8fae6b941-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xbsxf\" (UID: \"08d0a292-6e77-4197-9877-72e8fae6b941\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9lp5\" (UniqueName: \"kubernetes.io/projected/d9a53c5a-f3a7-484e-9058-11e61da9c9ed-kube-api-access-l9lp5\") pod \"openshift-config-operator-7777fb866f-mqxzb\" (UID: \"d9a53c5a-f3a7-484e-9058-11e61da9c9ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977257 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r46j\" (UniqueName: \"kubernetes.io/projected/fff4c3fa-a124-4cb5-806d-eea27e5dfd4a-kube-api-access-4r46j\") pod \"downloads-7954f5f757-jtxfb\" (UID: \"fff4c3fa-a124-4cb5-806d-eea27e5dfd4a\") " pod="openshift-console/downloads-7954f5f757-jtxfb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977290 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jwdb\" (UniqueName: \"kubernetes.io/projected/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-kube-api-access-2jwdb\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977306 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dccff42e-1a6b-4239-a095-35af760b79b1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7mdsb\" (UID: \"dccff42e-1a6b-4239-a095-35af760b79b1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977324 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9a53c5a-f3a7-484e-9058-11e61da9c9ed-serving-cert\") pod \"openshift-config-operator-7777fb866f-mqxzb\" (UID: \"d9a53c5a-f3a7-484e-9058-11e61da9c9ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977357 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-machine-approver-tls\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977383 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkr26\" (UniqueName: \"kubernetes.io/projected/02f0a77f-7099-480e-b4e6-6183592ca5e6-kube-api-access-hkr26\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d397f9a-119c-4d88-ae8d-68ed99c919f3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977429 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzf5d\" (UniqueName: \"kubernetes.io/projected/93429dc0-c4aa-4eec-a48d-7128e13d0d8f-kube-api-access-dzf5d\") pod \"openshift-controller-manager-operator-756b6f6bc6-kwxf5\" (UID: \"93429dc0-c4aa-4eec-a48d-7128e13d0d8f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977455 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dccff42e-1a6b-4239-a095-35af760b79b1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7mdsb\" (UID: \"dccff42e-1a6b-4239-a095-35af760b79b1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60414207-aa6c-42ec-a02c-562a10239881-config\") pod \"kube-controller-manager-operator-78b949d7b-kzwwp\" (UID: \"60414207-aa6c-42ec-a02c-562a10239881\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977494 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d94e93e8-457e-4b40-b09d-480ff253bb8a-trusted-ca\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/86022daa-0ebf-4820-a476-e5cacc43f9d5-metrics-tls\") pod \"dns-operator-744455d44c-gzfwc\" (UID: \"86022daa-0ebf-4820-a476-e5cacc43f9d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977525 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-oauth-serving-cert\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977541 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94e93e8-457e-4b40-b09d-480ff253bb8a-config\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02f0a77f-7099-480e-b4e6-6183592ca5e6-config\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977576 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9vkq\" (UniqueName: \"kubernetes.io/projected/68ba89bb-2ac5-4204-8362-891783709611-kube-api-access-x9vkq\") pod \"openshift-apiserver-operator-796bbdcf4f-hxvj9\" (UID: \"68ba89bb-2ac5-4204-8362-891783709611\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977592 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977609 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.977731 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-policies\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.978523 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ba89bb-2ac5-4204-8362-891783709611-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hxvj9\" (UID: \"68ba89bb-2ac5-4204-8362-891783709611\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.978585 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-dir\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.978983 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93429dc0-c4aa-4eec-a48d-7128e13d0d8f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kwxf5\" (UID: \"93429dc0-c4aa-4eec-a48d-7128e13d0d8f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.979783 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-auth-proxy-config\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.980018 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.980826 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-config\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.981455 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-client-ca\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.981481 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/02f0a77f-7099-480e-b4e6-6183592ca5e6-etcd-ca\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.981533 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-trusted-ca-bundle\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.981705 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.981806 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-oauth-serving-cert\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.981992 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-config\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.982474 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-images\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.982522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02f0a77f-7099-480e-b4e6-6183592ca5e6-serving-cert\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.982579 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-config\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.982960 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d94e93e8-457e-4b40-b09d-480ff253bb8a-trusted-ca\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.983061 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.983217 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d9a53c5a-f3a7-484e-9058-11e61da9c9ed-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mqxzb\" (UID: \"d9a53c5a-f3a7-484e-9058-11e61da9c9ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.984395 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-service-ca\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.984568 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d397f9a-119c-4d88-ae8d-68ed99c919f3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.984625 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94e93e8-457e-4b40-b09d-480ff253bb8a-config\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.985073 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-config\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.985618 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.985784 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-machine-approver-tls\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.986064 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.986515 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6349461-b5ae-4e1a-9711-9d676f1ec01f-serving-cert\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.986729 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s4lpp\" (UID: \"d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.987167 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.987261 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9a53c5a-f3a7-484e-9058-11e61da9c9ed-serving-cert\") pod \"openshift-config-operator-7777fb866f-mqxzb\" (UID: \"d9a53c5a-f3a7-484e-9058-11e61da9c9ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.987462 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.987490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.987580 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68ba89bb-2ac5-4204-8362-891783709611-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hxvj9\" (UID: \"68ba89bb-2ac5-4204-8362-891783709611\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.987681 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.988111 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.988671 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d94e93e8-457e-4b40-b09d-480ff253bb8a-serving-cert\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.988686 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.989263 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.990316 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2d397f9a-119c-4d88-ae8d-68ed99c919f3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.990411 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-oauth-config\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.990546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/86022daa-0ebf-4820-a476-e5cacc43f9d5-metrics-tls\") pod \"dns-operator-744455d44c-gzfwc\" (UID: \"86022daa-0ebf-4820-a476-e5cacc43f9d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.990854 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/02f0a77f-7099-480e-b4e6-6183592ca5e6-etcd-service-ca\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:37 crc kubenswrapper[4809]: I1127 17:11:37.991096 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-serving-cert\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.003553 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.003974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93429dc0-c4aa-4eec-a48d-7128e13d0d8f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kwxf5\" (UID: \"93429dc0-c4aa-4eec-a48d-7128e13d0d8f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.023872 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.033809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/02f0a77f-7099-480e-b4e6-6183592ca5e6-etcd-client\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.044234 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.053390 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02f0a77f-7099-480e-b4e6-6183592ca5e6-config\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.063663 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.084176 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.103603 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.123552 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.143850 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.153293 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60414207-aa6c-42ec-a02c-562a10239881-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kzwwp\" (UID: \"60414207-aa6c-42ec-a02c-562a10239881\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.164010 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.183424 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.192989 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60414207-aa6c-42ec-a02c-562a10239881-config\") pod \"kube-controller-manager-operator-78b949d7b-kzwwp\" (UID: \"60414207-aa6c-42ec-a02c-562a10239881\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.203623 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.224210 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.233514 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-metrics-tls\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.251411 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.262516 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-trusted-ca\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.263364 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.283550 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.302716 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.322887 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.343317 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.356285 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08d0a292-6e77-4197-9877-72e8fae6b941-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xbsxf\" (UID: \"08d0a292-6e77-4197-9877-72e8fae6b941\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.364267 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.375098 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d0a292-6e77-4197-9877-72e8fae6b941-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xbsxf\" (UID: \"08d0a292-6e77-4197-9877-72e8fae6b941\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.384881 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.404499 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.411203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dccff42e-1a6b-4239-a095-35af760b79b1-config\") pod \"kube-apiserver-operator-766d6c64bb-7mdsb\" (UID: \"dccff42e-1a6b-4239-a095-35af760b79b1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.424601 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.444720 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.454489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dccff42e-1a6b-4239-a095-35af760b79b1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7mdsb\" (UID: \"dccff42e-1a6b-4239-a095-35af760b79b1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.464638 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.524901 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.543973 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.564240 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.584365 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.604531 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.624908 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.644622 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.664585 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.684311 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.704153 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.725967 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.743713 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.765409 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.786008 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.804530 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.824201 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.843007 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.861779 4809 request.go:700] Waited for 1.008059252s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serviceaccount-dockercfg-rq7zk&limit=500&resourceVersion=0 Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.863466 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.884202 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.904073 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.923978 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.943874 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.964904 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 27 17:11:38 crc kubenswrapper[4809]: I1127 17:11:38.984160 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.004440 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.024265 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.044296 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.065526 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.083717 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.104137 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.124089 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.144905 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.165128 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.183977 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.223476 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.224329 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.244352 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.264550 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.284912 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.304798 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.324019 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.345168 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.365552 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.383916 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.403135 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.422947 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.445688 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.464457 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.484033 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.504270 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.524992 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.544207 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.583026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmbqq\" (UniqueName: \"kubernetes.io/projected/54856698-dafa-4315-84c3-e0b746e815f9-kube-api-access-vmbqq\") pod \"apiserver-7bbb656c7d-wpq4q\" (UID: \"54856698-dafa-4315-84c3-e0b746e815f9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.603490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54jb7\" (UniqueName: \"kubernetes.io/projected/cb38e0de-3758-4e46-beb3-cf27c411725e-kube-api-access-54jb7\") pod \"authentication-operator-69f744f599-6pm2s\" (UID: \"cb38e0de-3758-4e46-beb3-cf27c411725e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.604972 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.623806 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.643689 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.679313 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b5rd\" (UniqueName: \"kubernetes.io/projected/23af03b4-a427-4e08-848b-0561b9cc7364-kube-api-access-9b5rd\") pod \"controller-manager-879f6c89f-2cphh\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.699466 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxfwr\" (UniqueName: \"kubernetes.io/projected/a4dd8902-6cca-49cf-b678-6e6e0865f0dc-kube-api-access-cxfwr\") pod \"apiserver-76f77b778f-hgmrs\" (UID: \"a4dd8902-6cca-49cf-b678-6e6e0865f0dc\") " pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.703484 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.723607 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.743770 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.763444 4809 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.783041 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.802874 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.813900 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.841482 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbb4h\" (UniqueName: \"kubernetes.io/projected/fb3ec517-319b-4bd9-87b9-19ee21f410b6-kube-api-access-qbb4h\") pod \"oauth-openshift-558db77b4-2t2fb\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.858688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dptq\" (UniqueName: \"kubernetes.io/projected/08d0a292-6e77-4197-9877-72e8fae6b941-kube-api-access-2dptq\") pod \"kube-storage-version-migrator-operator-b67b599dd-xbsxf\" (UID: \"08d0a292-6e77-4197-9877-72e8fae6b941\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.862392 4809 request.go:700] Waited for 1.884195197s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.862923 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.878970 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.880508 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvtqt\" (UniqueName: \"kubernetes.io/projected/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-kube-api-access-kvtqt\") pod \"console-f9d7485db-w7f6q\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.894540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.904417 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbtm7\" (UniqueName: \"kubernetes.io/projected/6d8b54b5-a20b-4461-bca8-c8ac2ec82048-kube-api-access-sbtm7\") pod \"machine-api-operator-5694c8668f-6rxcc\" (UID: \"6d8b54b5-a20b-4461-bca8-c8ac2ec82048\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.926953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfjj4\" (UniqueName: \"kubernetes.io/projected/2d397f9a-119c-4d88-ae8d-68ed99c919f3-kube-api-access-tfjj4\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.937601 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.940763 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg7hg\" (UniqueName: \"kubernetes.io/projected/dcde5369-61c9-4e6a-9d06-66a8ba0337fc-kube-api-access-lg7hg\") pod \"machine-approver-56656f9798-l4tmc\" (UID: \"dcde5369-61c9-4e6a-9d06-66a8ba0337fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.951730 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.962856 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.971891 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpsk8\" (UniqueName: \"kubernetes.io/projected/d94e93e8-457e-4b40-b09d-480ff253bb8a-kube-api-access-mpsk8\") pod \"console-operator-58897d9998-7c7gv\" (UID: \"d94e93e8-457e-4b40-b09d-480ff253bb8a\") " pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:39 crc kubenswrapper[4809]: I1127 17:11:39.983566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r46j\" (UniqueName: \"kubernetes.io/projected/fff4c3fa-a124-4cb5-806d-eea27e5dfd4a-kube-api-access-4r46j\") pod \"downloads-7954f5f757-jtxfb\" (UID: \"fff4c3fa-a124-4cb5-806d-eea27e5dfd4a\") " pod="openshift-console/downloads-7954f5f757-jtxfb" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.001367 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwvlz\" (UniqueName: \"kubernetes.io/projected/f6349461-b5ae-4e1a-9711-9d676f1ec01f-kube-api-access-pwvlz\") pod \"route-controller-manager-6576b87f9c-wq2f9\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.005989 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.023279 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.030314 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dccff42e-1a6b-4239-a095-35af760b79b1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7mdsb\" (UID: \"dccff42e-1a6b-4239-a095-35af760b79b1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.030769 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.040641 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jtxfb" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.041300 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.060577 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jwdb\" (UniqueName: \"kubernetes.io/projected/be0ee2fc-9e5a-488f-8878-f29c0a7e6393-kube-api-access-2jwdb\") pod \"ingress-operator-5b745b69d9-ppfvt\" (UID: \"be0ee2fc-9e5a-488f-8878-f29c0a7e6393\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.063538 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.093638 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frzs8\" (UniqueName: \"kubernetes.io/projected/d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4-kube-api-access-frzs8\") pod \"cluster-samples-operator-665b6dd947-s4lpp\" (UID: \"d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.109348 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2cphh"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.111649 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6pm2s"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.119956 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60414207-aa6c-42ec-a02c-562a10239881-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kzwwp\" (UID: \"60414207-aa6c-42ec-a02c-562a10239881\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.135675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkr26\" (UniqueName: \"kubernetes.io/projected/02f0a77f-7099-480e-b4e6-6183592ca5e6-kube-api-access-hkr26\") pod \"etcd-operator-b45778765-m5lpz\" (UID: \"02f0a77f-7099-480e-b4e6-6183592ca5e6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.136099 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.144216 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.149069 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.159500 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.164408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d397f9a-119c-4d88-ae8d-68ed99c919f3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lzk7p\" (UID: \"2d397f9a-119c-4d88-ae8d-68ed99c919f3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.172136 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" event={"ID":"dcde5369-61c9-4e6a-9d06-66a8ba0337fc","Type":"ContainerStarted","Data":"c3fc44b2c96c433cc7ed97ad455eb32b8736118ba2789998bc8e8e35ea8273a7"} Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.174194 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" event={"ID":"54856698-dafa-4315-84c3-e0b746e815f9","Type":"ContainerStarted","Data":"7b7f262f1b586d7df3e8625710c9137fbf0f37aaffafc3b7d7071291532ca185"} Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.176132 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzf5d\" (UniqueName: \"kubernetes.io/projected/93429dc0-c4aa-4eec-a48d-7128e13d0d8f-kube-api-access-dzf5d\") pod \"openshift-controller-manager-operator-756b6f6bc6-kwxf5\" (UID: \"93429dc0-c4aa-4eec-a48d-7128e13d0d8f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.185971 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llrhw\" (UniqueName: \"kubernetes.io/projected/86022daa-0ebf-4820-a476-e5cacc43f9d5-kube-api-access-llrhw\") pod \"dns-operator-744455d44c-gzfwc\" (UID: \"86022daa-0ebf-4820-a476-e5cacc43f9d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.198481 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9lp5\" (UniqueName: \"kubernetes.io/projected/d9a53c5a-f3a7-484e-9058-11e61da9c9ed-kube-api-access-l9lp5\") pod \"openshift-config-operator-7777fb866f-mqxzb\" (UID: \"d9a53c5a-f3a7-484e-9058-11e61da9c9ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:40 crc kubenswrapper[4809]: W1127 17:11:40.208038 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23af03b4_a427_4e08_848b_0561b9cc7364.slice/crio-dac87afede946ab918c20020e54ac5680187e2ff813db1db1e915627022993e1 WatchSource:0}: Error finding container dac87afede946ab918c20020e54ac5680187e2ff813db1db1e915627022993e1: Status 404 returned error can't find the container with id dac87afede946ab918c20020e54ac5680187e2ff813db1db1e915627022993e1 Nov 27 17:11:40 crc kubenswrapper[4809]: W1127 17:11:40.209659 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb38e0de_3758_4e46_beb3_cf27c411725e.slice/crio-d7800dcc5f47643b27cbaff7031aafe192abd30f98d0182a1f926871cf8afe46 WatchSource:0}: Error finding container d7800dcc5f47643b27cbaff7031aafe192abd30f98d0182a1f926871cf8afe46: Status 404 returned error can't find the container with id d7800dcc5f47643b27cbaff7031aafe192abd30f98d0182a1f926871cf8afe46 Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.216937 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9vkq\" (UniqueName: \"kubernetes.io/projected/68ba89bb-2ac5-4204-8362-891783709611-kube-api-access-x9vkq\") pod \"openshift-apiserver-operator-796bbdcf4f-hxvj9\" (UID: \"68ba89bb-2ac5-4204-8362-891783709611\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.274044 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.278078 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hgmrs"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.313898 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-registry-certificates\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.313933 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-registry-tls\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.313955 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/480a6028-0611-40c1-8204-0e2e37608800-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.314016 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-trusted-ca\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.314036 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/480a6028-0611-40c1-8204-0e2e37608800-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.314073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.314091 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x47mb\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-kube-api-access-x47mb\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.314129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-bound-sa-token\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.314508 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:40.814494979 +0000 UTC m=+136.086952331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.314759 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.315162 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6rxcc"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.348602 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.359978 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.361150 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-w7f6q"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.367731 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.374173 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.382328 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.416454 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.416639 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:40.916614434 +0000 UTC m=+136.189071786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.418045 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-trusted-ca\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.418115 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw6ng\" (UniqueName: \"kubernetes.io/projected/cbb30a57-2b12-464b-b9aa-6a54d8314abb-kube-api-access-cw6ng\") pod \"olm-operator-6b444d44fb-qt7fm\" (UID: \"cbb30a57-2b12-464b-b9aa-6a54d8314abb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.418152 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-service-ca-bundle\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.418213 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-94nbr\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.418265 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0a9bd25f-e243-4126-bddd-2a9478e35618-node-bootstrap-token\") pod \"machine-config-server-zpwvm\" (UID: \"0a9bd25f-e243-4126-bddd-2a9478e35618\") " pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.418879 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa01a74-34e4-4ef5-b041-17d548feb09f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6l2d\" (UID: \"8aa01a74-34e4-4ef5-b041-17d548feb09f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.418975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlswd\" (UniqueName: \"kubernetes.io/projected/29b8f288-c15b-4bfa-88fd-b7189659d57a-kube-api-access-nlswd\") pod \"service-ca-9c57cc56f-8lfbq\" (UID: \"29b8f288-c15b-4bfa-88fd-b7189659d57a\") " pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.419249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.419403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x47mb\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-kube-api-access-x47mb\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.419607 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrtz6\" (UniqueName: \"kubernetes.io/projected/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-kube-api-access-zrtz6\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.419688 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:40.919669152 +0000 UTC m=+136.192126494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.419733 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwfzs\" (UniqueName: \"kubernetes.io/projected/f62405dd-1836-4b7d-a785-27f34188a241-kube-api-access-vwfzs\") pod \"service-ca-operator-777779d784-zvlsm\" (UID: \"f62405dd-1836-4b7d-a785-27f34188a241\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.420062 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1793bd02-489b-422c-9541-4354b4c4c020-srv-cert\") pod \"catalog-operator-68c6474976-j6ggx\" (UID: \"1793bd02-489b-422c-9541-4354b4c4c020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.420121 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds4tj\" (UniqueName: \"kubernetes.io/projected/f6628245-a537-4cf1-8b4c-207bb7cb3d0b-kube-api-access-ds4tj\") pod \"dns-default-6v9pw\" (UID: \"f6628245-a537-4cf1-8b4c-207bb7cb3d0b\") " pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.420152 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f62405dd-1836-4b7d-a785-27f34188a241-serving-cert\") pod \"service-ca-operator-777779d784-zvlsm\" (UID: \"f62405dd-1836-4b7d-a785-27f34188a241\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.420189 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/455e5a5e-24c5-4a05-944d-d15d8b3c2b1b-proxy-tls\") pod \"machine-config-controller-84d6567774-glrj4\" (UID: \"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.420224 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-94nbr\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.420249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwq7z\" (UniqueName: \"kubernetes.io/projected/455e5a5e-24c5-4a05-944d-d15d8b3c2b1b-kube-api-access-dwq7z\") pod \"machine-config-controller-84d6567774-glrj4\" (UID: \"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.420288 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-trusted-ca\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.420355 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/455e5a5e-24c5-4a05-944d-d15d8b3c2b1b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-glrj4\" (UID: \"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.420677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b9e66d8f-4975-4d8c-873d-4c1e683044df-proxy-tls\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.421145 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhbmt\" (UniqueName: \"kubernetes.io/projected/92093439-c71b-4fdd-88ff-189ecd253269-kube-api-access-zhbmt\") pod \"marketplace-operator-79b997595-94nbr\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.421214 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtn56\" (UniqueName: \"kubernetes.io/projected/87e2c04c-53e6-4524-90b8-b122e31b9280-kube-api-access-gtn56\") pod \"control-plane-machine-set-operator-78cbb6b69f-tvl4x\" (UID: \"87e2c04c-53e6-4524-90b8-b122e31b9280\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.421233 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-mountpoint-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.421613 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b584acf2-534c-4ce1-999e-3d77468b9bd7-apiservice-cert\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.421658 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cbb30a57-2b12-464b-b9aa-6a54d8314abb-srv-cert\") pod \"olm-operator-6b444d44fb-qt7fm\" (UID: \"cbb30a57-2b12-464b-b9aa-6a54d8314abb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.421687 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8aa01a74-34e4-4ef5-b041-17d548feb09f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6l2d\" (UID: \"8aa01a74-34e4-4ef5-b041-17d548feb09f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.421723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/480a6028-0611-40c1-8204-0e2e37608800-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.422013 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/480a6028-0611-40c1-8204-0e2e37608800-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.422033 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-metrics-certs\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.422166 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b584acf2-534c-4ce1-999e-3d77468b9bd7-tmpfs\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.422239 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lndlc\" (UniqueName: \"kubernetes.io/projected/1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8-kube-api-access-lndlc\") pod \"multus-admission-controller-857f4d67dd-mcsp5\" (UID: \"1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.422322 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b9e66d8f-4975-4d8c-873d-4c1e683044df-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.422544 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmgrc\" (UniqueName: \"kubernetes.io/projected/73315a03-8ebc-44c1-bc28-0fa0965b1753-kube-api-access-bmgrc\") pod \"ingress-canary-r7lqs\" (UID: \"73315a03-8ebc-44c1-bc28-0fa0965b1753\") " pod="openshift-ingress-canary/ingress-canary-r7lqs" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.422588 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8cvd\" (UniqueName: \"kubernetes.io/projected/4f877f18-1c0c-43b7-9873-9bd3d48012d4-kube-api-access-p8cvd\") pod \"package-server-manager-789f6589d5-rsqzp\" (UID: \"4f877f18-1c0c-43b7-9873-9bd3d48012d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.422945 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mcsp5\" (UID: \"1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.423014 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftzhv\" (UniqueName: \"kubernetes.io/projected/0a9bd25f-e243-4126-bddd-2a9478e35618-kube-api-access-ftzhv\") pod \"machine-config-server-zpwvm\" (UID: \"0a9bd25f-e243-4126-bddd-2a9478e35618\") " pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.423148 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wmgx\" (UniqueName: \"kubernetes.io/projected/b9e66d8f-4975-4d8c-873d-4c1e683044df-kube-api-access-4wmgx\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.423218 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/480a6028-0611-40c1-8204-0e2e37608800-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.423267 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv98z\" (UniqueName: \"kubernetes.io/projected/15c3167d-77d3-457f-b602-5e87f96afe8d-kube-api-access-tv98z\") pod \"collect-profiles-29404380-j6wjd\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.423523 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15c3167d-77d3-457f-b602-5e87f96afe8d-config-volume\") pod \"collect-profiles-29404380-j6wjd\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.423570 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6628245-a537-4cf1-8b4c-207bb7cb3d0b-config-volume\") pod \"dns-default-6v9pw\" (UID: \"f6628245-a537-4cf1-8b4c-207bb7cb3d0b\") " pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.424085 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-csi-data-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.424442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqlzm\" (UniqueName: \"kubernetes.io/projected/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-kube-api-access-gqlzm\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.424468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/29b8f288-c15b-4bfa-88fd-b7189659d57a-signing-cabundle\") pod \"service-ca-9c57cc56f-8lfbq\" (UID: \"29b8f288-c15b-4bfa-88fd-b7189659d57a\") " pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.424602 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/29b8f288-c15b-4bfa-88fd-b7189659d57a-signing-key\") pod \"service-ca-9c57cc56f-8lfbq\" (UID: \"29b8f288-c15b-4bfa-88fd-b7189659d57a\") " pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.424823 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-socket-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.425146 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-bound-sa-token\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.425184 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f62405dd-1836-4b7d-a785-27f34188a241-config\") pod \"service-ca-operator-777779d784-zvlsm\" (UID: \"f62405dd-1836-4b7d-a785-27f34188a241\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.425224 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cbb30a57-2b12-464b-b9aa-6a54d8314abb-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qt7fm\" (UID: \"cbb30a57-2b12-464b-b9aa-6a54d8314abb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.425254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-plugins-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.425590 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8aa01a74-34e4-4ef5-b041-17d548feb09f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6l2d\" (UID: \"8aa01a74-34e4-4ef5-b041-17d548feb09f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.425633 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f877f18-1c0c-43b7-9873-9bd3d48012d4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rsqzp\" (UID: \"4f877f18-1c0c-43b7-9873-9bd3d48012d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.425670 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-default-certificate\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.426007 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxql6\" (UniqueName: \"kubernetes.io/projected/132b91ef-f529-4194-b6cd-f356fc4d0e33-kube-api-access-mxql6\") pod \"migrator-59844c95c7-tf68p\" (UID: \"132b91ef-f529-4194-b6cd-f356fc4d0e33\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.426192 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15c3167d-77d3-457f-b602-5e87f96afe8d-secret-volume\") pod \"collect-profiles-29404380-j6wjd\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427170 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-stats-auth\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427285 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-registry-certificates\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427346 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-registry-tls\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73315a03-8ebc-44c1-bc28-0fa0965b1753-cert\") pod \"ingress-canary-r7lqs\" (UID: \"73315a03-8ebc-44c1-bc28-0fa0965b1753\") " pod="openshift-ingress-canary/ingress-canary-r7lqs" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427400 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm2cl\" (UniqueName: \"kubernetes.io/projected/b584acf2-534c-4ce1-999e-3d77468b9bd7-kube-api-access-qm2cl\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427418 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f6628245-a537-4cf1-8b4c-207bb7cb3d0b-metrics-tls\") pod \"dns-default-6v9pw\" (UID: \"f6628245-a537-4cf1-8b4c-207bb7cb3d0b\") " pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427466 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0a9bd25f-e243-4126-bddd-2a9478e35618-certs\") pod \"machine-config-server-zpwvm\" (UID: \"0a9bd25f-e243-4126-bddd-2a9478e35618\") " pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427488 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1793bd02-489b-422c-9541-4354b4c4c020-profile-collector-cert\") pod \"catalog-operator-68c6474976-j6ggx\" (UID: \"1793bd02-489b-422c-9541-4354b4c4c020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427523 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b584acf2-534c-4ce1-999e-3d77468b9bd7-webhook-cert\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwspf\" (UniqueName: \"kubernetes.io/projected/1793bd02-489b-422c-9541-4354b4c4c020-kube-api-access-xwspf\") pod \"catalog-operator-68c6474976-j6ggx\" (UID: \"1793bd02-489b-422c-9541-4354b4c4c020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427562 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-registration-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427593 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/87e2c04c-53e6-4524-90b8-b122e31b9280-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-tvl4x\" (UID: \"87e2c04c-53e6-4524-90b8-b122e31b9280\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.427645 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b9e66d8f-4975-4d8c-873d-4c1e683044df-images\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.429186 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-registry-certificates\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.441621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-registry-tls\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.446108 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/480a6028-0611-40c1-8204-0e2e37608800-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.468247 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x47mb\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-kube-api-access-x47mb\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.483660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-bound-sa-token\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.528383 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.528571 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.028543743 +0000 UTC m=+136.301001095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.528921 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw6ng\" (UniqueName: \"kubernetes.io/projected/cbb30a57-2b12-464b-b9aa-6a54d8314abb-kube-api-access-cw6ng\") pod \"olm-operator-6b444d44fb-qt7fm\" (UID: \"cbb30a57-2b12-464b-b9aa-6a54d8314abb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.528950 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-service-ca-bundle\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.528967 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-94nbr\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.528985 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0a9bd25f-e243-4126-bddd-2a9478e35618-node-bootstrap-token\") pod \"machine-config-server-zpwvm\" (UID: \"0a9bd25f-e243-4126-bddd-2a9478e35618\") " pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529007 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa01a74-34e4-4ef5-b041-17d548feb09f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6l2d\" (UID: \"8aa01a74-34e4-4ef5-b041-17d548feb09f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529027 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlswd\" (UniqueName: \"kubernetes.io/projected/29b8f288-c15b-4bfa-88fd-b7189659d57a-kube-api-access-nlswd\") pod \"service-ca-9c57cc56f-8lfbq\" (UID: \"29b8f288-c15b-4bfa-88fd-b7189659d57a\") " pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529050 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529068 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrtz6\" (UniqueName: \"kubernetes.io/projected/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-kube-api-access-zrtz6\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529088 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwfzs\" (UniqueName: \"kubernetes.io/projected/f62405dd-1836-4b7d-a785-27f34188a241-kube-api-access-vwfzs\") pod \"service-ca-operator-777779d784-zvlsm\" (UID: \"f62405dd-1836-4b7d-a785-27f34188a241\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1793bd02-489b-422c-9541-4354b4c4c020-srv-cert\") pod \"catalog-operator-68c6474976-j6ggx\" (UID: \"1793bd02-489b-422c-9541-4354b4c4c020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529130 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds4tj\" (UniqueName: \"kubernetes.io/projected/f6628245-a537-4cf1-8b4c-207bb7cb3d0b-kube-api-access-ds4tj\") pod \"dns-default-6v9pw\" (UID: \"f6628245-a537-4cf1-8b4c-207bb7cb3d0b\") " pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529145 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f62405dd-1836-4b7d-a785-27f34188a241-serving-cert\") pod \"service-ca-operator-777779d784-zvlsm\" (UID: \"f62405dd-1836-4b7d-a785-27f34188a241\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529160 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/455e5a5e-24c5-4a05-944d-d15d8b3c2b1b-proxy-tls\") pod \"machine-config-controller-84d6567774-glrj4\" (UID: \"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529177 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-94nbr\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529194 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwq7z\" (UniqueName: \"kubernetes.io/projected/455e5a5e-24c5-4a05-944d-d15d8b3c2b1b-kube-api-access-dwq7z\") pod \"machine-config-controller-84d6567774-glrj4\" (UID: \"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/455e5a5e-24c5-4a05-944d-d15d8b3c2b1b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-glrj4\" (UID: \"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529229 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b9e66d8f-4975-4d8c-873d-4c1e683044df-proxy-tls\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529253 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhbmt\" (UniqueName: \"kubernetes.io/projected/92093439-c71b-4fdd-88ff-189ecd253269-kube-api-access-zhbmt\") pod \"marketplace-operator-79b997595-94nbr\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529272 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtn56\" (UniqueName: \"kubernetes.io/projected/87e2c04c-53e6-4524-90b8-b122e31b9280-kube-api-access-gtn56\") pod \"control-plane-machine-set-operator-78cbb6b69f-tvl4x\" (UID: \"87e2c04c-53e6-4524-90b8-b122e31b9280\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529287 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-mountpoint-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529302 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cbb30a57-2b12-464b-b9aa-6a54d8314abb-srv-cert\") pod \"olm-operator-6b444d44fb-qt7fm\" (UID: \"cbb30a57-2b12-464b-b9aa-6a54d8314abb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b584acf2-534c-4ce1-999e-3d77468b9bd7-apiservice-cert\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529333 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8aa01a74-34e4-4ef5-b041-17d548feb09f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6l2d\" (UID: \"8aa01a74-34e4-4ef5-b041-17d548feb09f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529354 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-metrics-certs\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529372 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lndlc\" (UniqueName: \"kubernetes.io/projected/1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8-kube-api-access-lndlc\") pod \"multus-admission-controller-857f4d67dd-mcsp5\" (UID: \"1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b584acf2-534c-4ce1-999e-3d77468b9bd7-tmpfs\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b9e66d8f-4975-4d8c-873d-4c1e683044df-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529430 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmgrc\" (UniqueName: \"kubernetes.io/projected/73315a03-8ebc-44c1-bc28-0fa0965b1753-kube-api-access-bmgrc\") pod \"ingress-canary-r7lqs\" (UID: \"73315a03-8ebc-44c1-bc28-0fa0965b1753\") " pod="openshift-ingress-canary/ingress-canary-r7lqs" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529444 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8cvd\" (UniqueName: \"kubernetes.io/projected/4f877f18-1c0c-43b7-9873-9bd3d48012d4-kube-api-access-p8cvd\") pod \"package-server-manager-789f6589d5-rsqzp\" (UID: \"4f877f18-1c0c-43b7-9873-9bd3d48012d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529461 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mcsp5\" (UID: \"1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529476 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftzhv\" (UniqueName: \"kubernetes.io/projected/0a9bd25f-e243-4126-bddd-2a9478e35618-kube-api-access-ftzhv\") pod \"machine-config-server-zpwvm\" (UID: \"0a9bd25f-e243-4126-bddd-2a9478e35618\") " pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529492 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wmgx\" (UniqueName: \"kubernetes.io/projected/b9e66d8f-4975-4d8c-873d-4c1e683044df-kube-api-access-4wmgx\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529509 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv98z\" (UniqueName: \"kubernetes.io/projected/15c3167d-77d3-457f-b602-5e87f96afe8d-kube-api-access-tv98z\") pod \"collect-profiles-29404380-j6wjd\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529525 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15c3167d-77d3-457f-b602-5e87f96afe8d-config-volume\") pod \"collect-profiles-29404380-j6wjd\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529539 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6628245-a537-4cf1-8b4c-207bb7cb3d0b-config-volume\") pod \"dns-default-6v9pw\" (UID: \"f6628245-a537-4cf1-8b4c-207bb7cb3d0b\") " pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-csi-data-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529573 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqlzm\" (UniqueName: \"kubernetes.io/projected/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-kube-api-access-gqlzm\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/29b8f288-c15b-4bfa-88fd-b7189659d57a-signing-cabundle\") pod \"service-ca-9c57cc56f-8lfbq\" (UID: \"29b8f288-c15b-4bfa-88fd-b7189659d57a\") " pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529603 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/29b8f288-c15b-4bfa-88fd-b7189659d57a-signing-key\") pod \"service-ca-9c57cc56f-8lfbq\" (UID: \"29b8f288-c15b-4bfa-88fd-b7189659d57a\") " pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-socket-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529633 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cbb30a57-2b12-464b-b9aa-6a54d8314abb-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qt7fm\" (UID: \"cbb30a57-2b12-464b-b9aa-6a54d8314abb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f62405dd-1836-4b7d-a785-27f34188a241-config\") pod \"service-ca-operator-777779d784-zvlsm\" (UID: \"f62405dd-1836-4b7d-a785-27f34188a241\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529879 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-plugins-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529903 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8aa01a74-34e4-4ef5-b041-17d548feb09f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6l2d\" (UID: \"8aa01a74-34e4-4ef5-b041-17d548feb09f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529926 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f877f18-1c0c-43b7-9873-9bd3d48012d4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rsqzp\" (UID: \"4f877f18-1c0c-43b7-9873-9bd3d48012d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529946 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-default-certificate\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529968 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxql6\" (UniqueName: \"kubernetes.io/projected/132b91ef-f529-4194-b6cd-f356fc4d0e33-kube-api-access-mxql6\") pod \"migrator-59844c95c7-tf68p\" (UID: \"132b91ef-f529-4194-b6cd-f356fc4d0e33\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.529988 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15c3167d-77d3-457f-b602-5e87f96afe8d-secret-volume\") pod \"collect-profiles-29404380-j6wjd\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530007 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-stats-auth\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530033 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73315a03-8ebc-44c1-bc28-0fa0965b1753-cert\") pod \"ingress-canary-r7lqs\" (UID: \"73315a03-8ebc-44c1-bc28-0fa0965b1753\") " pod="openshift-ingress-canary/ingress-canary-r7lqs" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm2cl\" (UniqueName: \"kubernetes.io/projected/b584acf2-534c-4ce1-999e-3d77468b9bd7-kube-api-access-qm2cl\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f6628245-a537-4cf1-8b4c-207bb7cb3d0b-metrics-tls\") pod \"dns-default-6v9pw\" (UID: \"f6628245-a537-4cf1-8b4c-207bb7cb3d0b\") " pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530101 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0a9bd25f-e243-4126-bddd-2a9478e35618-certs\") pod \"machine-config-server-zpwvm\" (UID: \"0a9bd25f-e243-4126-bddd-2a9478e35618\") " pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530123 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1793bd02-489b-422c-9541-4354b4c4c020-profile-collector-cert\") pod \"catalog-operator-68c6474976-j6ggx\" (UID: \"1793bd02-489b-422c-9541-4354b4c4c020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwspf\" (UniqueName: \"kubernetes.io/projected/1793bd02-489b-422c-9541-4354b4c4c020-kube-api-access-xwspf\") pod \"catalog-operator-68c6474976-j6ggx\" (UID: \"1793bd02-489b-422c-9541-4354b4c4c020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530164 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-registration-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b584acf2-534c-4ce1-999e-3d77468b9bd7-webhook-cert\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530208 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/87e2c04c-53e6-4524-90b8-b122e31b9280-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-tvl4x\" (UID: \"87e2c04c-53e6-4524-90b8-b122e31b9280\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.530238 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b9e66d8f-4975-4d8c-873d-4c1e683044df-images\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.531122 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.031102077 +0000 UTC m=+136.303559499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.531225 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b9e66d8f-4975-4d8c-873d-4c1e683044df-images\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.531279 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-csi-data-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.532451 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6628245-a537-4cf1-8b4c-207bb7cb3d0b-config-volume\") pod \"dns-default-6v9pw\" (UID: \"f6628245-a537-4cf1-8b4c-207bb7cb3d0b\") " pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.532449 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-registration-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.532622 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f62405dd-1836-4b7d-a785-27f34188a241-config\") pod \"service-ca-operator-777779d784-zvlsm\" (UID: \"f62405dd-1836-4b7d-a785-27f34188a241\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.532842 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/29b8f288-c15b-4bfa-88fd-b7189659d57a-signing-cabundle\") pod \"service-ca-9c57cc56f-8lfbq\" (UID: \"29b8f288-c15b-4bfa-88fd-b7189659d57a\") " pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.532900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-plugins-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.536134 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-socket-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.536774 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-service-ca-bundle\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.537505 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-metrics-certs\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.537901 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-94nbr\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.538122 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b584acf2-534c-4ce1-999e-3d77468b9bd7-tmpfs\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.538281 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa01a74-34e4-4ef5-b041-17d548feb09f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6l2d\" (UID: \"8aa01a74-34e4-4ef5-b041-17d548feb09f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.538606 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b9e66d8f-4975-4d8c-873d-4c1e683044df-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.545155 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15c3167d-77d3-457f-b602-5e87f96afe8d-secret-volume\") pod \"collect-profiles-29404380-j6wjd\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.550696 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-mountpoint-dir\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.552879 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/455e5a5e-24c5-4a05-944d-d15d8b3c2b1b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-glrj4\" (UID: \"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.557025 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1793bd02-489b-422c-9541-4354b4c4c020-srv-cert\") pod \"catalog-operator-68c6474976-j6ggx\" (UID: \"1793bd02-489b-422c-9541-4354b4c4c020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.557275 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8aa01a74-34e4-4ef5-b041-17d548feb09f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6l2d\" (UID: \"8aa01a74-34e4-4ef5-b041-17d548feb09f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.557465 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f62405dd-1836-4b7d-a785-27f34188a241-serving-cert\") pod \"service-ca-operator-777779d784-zvlsm\" (UID: \"f62405dd-1836-4b7d-a785-27f34188a241\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.557962 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/29b8f288-c15b-4bfa-88fd-b7189659d57a-signing-key\") pod \"service-ca-9c57cc56f-8lfbq\" (UID: \"29b8f288-c15b-4bfa-88fd-b7189659d57a\") " pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.558011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0a9bd25f-e243-4126-bddd-2a9478e35618-node-bootstrap-token\") pod \"machine-config-server-zpwvm\" (UID: \"0a9bd25f-e243-4126-bddd-2a9478e35618\") " pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.558378 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cbb30a57-2b12-464b-b9aa-6a54d8314abb-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qt7fm\" (UID: \"cbb30a57-2b12-464b-b9aa-6a54d8314abb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.558540 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f877f18-1c0c-43b7-9873-9bd3d48012d4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rsqzp\" (UID: \"4f877f18-1c0c-43b7-9873-9bd3d48012d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.559161 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0a9bd25f-e243-4126-bddd-2a9478e35618-certs\") pod \"machine-config-server-zpwvm\" (UID: \"0a9bd25f-e243-4126-bddd-2a9478e35618\") " pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.559218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1793bd02-489b-422c-9541-4354b4c4c020-profile-collector-cert\") pod \"catalog-operator-68c6474976-j6ggx\" (UID: \"1793bd02-489b-422c-9541-4354b4c4c020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.559403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mcsp5\" (UID: \"1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.559443 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-default-certificate\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.559460 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b9e66d8f-4975-4d8c-873d-4c1e683044df-proxy-tls\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.559782 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-stats-auth\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.559926 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f6628245-a537-4cf1-8b4c-207bb7cb3d0b-metrics-tls\") pod \"dns-default-6v9pw\" (UID: \"f6628245-a537-4cf1-8b4c-207bb7cb3d0b\") " pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.560128 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b584acf2-534c-4ce1-999e-3d77468b9bd7-webhook-cert\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.560160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/87e2c04c-53e6-4524-90b8-b122e31b9280-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-tvl4x\" (UID: \"87e2c04c-53e6-4524-90b8-b122e31b9280\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.569118 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-94nbr\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.577023 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b584acf2-534c-4ce1-999e-3d77468b9bd7-apiservice-cert\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.579473 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15c3167d-77d3-457f-b602-5e87f96afe8d-config-volume\") pod \"collect-profiles-29404380-j6wjd\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.581299 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/455e5a5e-24c5-4a05-944d-d15d8b3c2b1b-proxy-tls\") pod \"machine-config-controller-84d6567774-glrj4\" (UID: \"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.584004 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cbb30a57-2b12-464b-b9aa-6a54d8314abb-srv-cert\") pod \"olm-operator-6b444d44fb-qt7fm\" (UID: \"cbb30a57-2b12-464b-b9aa-6a54d8314abb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.584015 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds4tj\" (UniqueName: \"kubernetes.io/projected/f6628245-a537-4cf1-8b4c-207bb7cb3d0b-kube-api-access-ds4tj\") pod \"dns-default-6v9pw\" (UID: \"f6628245-a537-4cf1-8b4c-207bb7cb3d0b\") " pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.584488 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73315a03-8ebc-44c1-bc28-0fa0965b1753-cert\") pod \"ingress-canary-r7lqs\" (UID: \"73315a03-8ebc-44c1-bc28-0fa0965b1753\") " pod="openshift-ingress-canary/ingress-canary-r7lqs" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.605138 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlswd\" (UniqueName: \"kubernetes.io/projected/29b8f288-c15b-4bfa-88fd-b7189659d57a-kube-api-access-nlswd\") pod \"service-ca-9c57cc56f-8lfbq\" (UID: \"29b8f288-c15b-4bfa-88fd-b7189659d57a\") " pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.622471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqlzm\" (UniqueName: \"kubernetes.io/projected/9a3cce9e-247e-457d-b41c-a1a7ee3c7888-kube-api-access-gqlzm\") pod \"router-default-5444994796-kzdxz\" (UID: \"9a3cce9e-247e-457d-b41c-a1a7ee3c7888\") " pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.625188 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.626543 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxql6\" (UniqueName: \"kubernetes.io/projected/132b91ef-f529-4194-b6cd-f356fc4d0e33-kube-api-access-mxql6\") pod \"migrator-59844c95c7-tf68p\" (UID: \"132b91ef-f529-4194-b6cd-f356fc4d0e33\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.632945 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.633970 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.133919353 +0000 UTC m=+136.406376715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.634201 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.634864 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.13485553 +0000 UTC m=+136.407312872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.663123 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8cvd\" (UniqueName: \"kubernetes.io/projected/4f877f18-1c0c-43b7-9873-9bd3d48012d4-kube-api-access-p8cvd\") pod \"package-server-manager-789f6589d5-rsqzp\" (UID: \"4f877f18-1c0c-43b7-9873-9bd3d48012d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.679827 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw6ng\" (UniqueName: \"kubernetes.io/projected/cbb30a57-2b12-464b-b9aa-6a54d8314abb-kube-api-access-cw6ng\") pod \"olm-operator-6b444d44fb-qt7fm\" (UID: \"cbb30a57-2b12-464b-b9aa-6a54d8314abb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.703937 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrtz6\" (UniqueName: \"kubernetes.io/projected/d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5-kube-api-access-zrtz6\") pod \"csi-hostpathplugin-dpfxv\" (UID: \"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5\") " pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.732154 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwfzs\" (UniqueName: \"kubernetes.io/projected/f62405dd-1836-4b7d-a785-27f34188a241-kube-api-access-vwfzs\") pod \"service-ca-operator-777779d784-zvlsm\" (UID: \"f62405dd-1836-4b7d-a785-27f34188a241\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.735061 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.735490 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.235462512 +0000 UTC m=+136.507919864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.756509 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.759121 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwspf\" (UniqueName: \"kubernetes.io/projected/1793bd02-489b-422c-9541-4354b4c4c020-kube-api-access-xwspf\") pod \"catalog-operator-68c6474976-j6ggx\" (UID: \"1793bd02-489b-422c-9541-4354b4c4c020\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.769139 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lndlc\" (UniqueName: \"kubernetes.io/projected/1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8-kube-api-access-lndlc\") pod \"multus-admission-controller-857f4d67dd-mcsp5\" (UID: \"1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" Nov 27 17:11:40 crc kubenswrapper[4809]: W1127 17:11:40.776753 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddccff42e_1a6b_4239_a095_35af760b79b1.slice/crio-a166877d489774d5450e4d8ae30b7f47d756d3f4206c561d19e2bc4e2e620aec WatchSource:0}: Error finding container a166877d489774d5450e4d8ae30b7f47d756d3f4206c561d19e2bc4e2e620aec: Status 404 returned error can't find the container with id a166877d489774d5450e4d8ae30b7f47d756d3f4206c561d19e2bc4e2e620aec Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.781482 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.785522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmgrc\" (UniqueName: \"kubernetes.io/projected/73315a03-8ebc-44c1-bc28-0fa0965b1753-kube-api-access-bmgrc\") pod \"ingress-canary-r7lqs\" (UID: \"73315a03-8ebc-44c1-bc28-0fa0965b1753\") " pod="openshift-ingress-canary/ingress-canary-r7lqs" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.788169 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.802623 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.804771 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm2cl\" (UniqueName: \"kubernetes.io/projected/b584acf2-534c-4ce1-999e-3d77468b9bd7-kube-api-access-qm2cl\") pod \"packageserver-d55dfcdfc-vjg9h\" (UID: \"b584acf2-534c-4ce1-999e-3d77468b9bd7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.806883 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.811953 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.827108 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.829319 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.835264 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.836410 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.836909 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.336895548 +0000 UTC m=+136.609352900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.843859 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.853086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wmgx\" (UniqueName: \"kubernetes.io/projected/b9e66d8f-4975-4d8c-873d-4c1e683044df-kube-api-access-4wmgx\") pod \"machine-config-operator-74547568cd-ngdqp\" (UID: \"b9e66d8f-4975-4d8c-873d-4c1e683044df\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.855660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftzhv\" (UniqueName: \"kubernetes.io/projected/0a9bd25f-e243-4126-bddd-2a9478e35618-kube-api-access-ftzhv\") pod \"machine-config-server-zpwvm\" (UID: \"0a9bd25f-e243-4126-bddd-2a9478e35618\") " pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.855878 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.867382 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.867385 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.872432 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv98z\" (UniqueName: \"kubernetes.io/projected/15c3167d-77d3-457f-b602-5e87f96afe8d-kube-api-access-tv98z\") pod \"collect-profiles-29404380-j6wjd\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.875426 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-r7lqs" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.876633 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jtxfb"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.878545 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7c7gv"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.880521 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2t2fb"] Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.883723 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.885514 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhbmt\" (UniqueName: \"kubernetes.io/projected/92093439-c71b-4fdd-88ff-189ecd253269-kube-api-access-zhbmt\") pod \"marketplace-operator-79b997595-94nbr\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.888922 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zpwvm" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.901519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtn56\" (UniqueName: \"kubernetes.io/projected/87e2c04c-53e6-4524-90b8-b122e31b9280-kube-api-access-gtn56\") pod \"control-plane-machine-set-operator-78cbb6b69f-tvl4x\" (UID: \"87e2c04c-53e6-4524-90b8-b122e31b9280\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.913526 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.919628 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwq7z\" (UniqueName: \"kubernetes.io/projected/455e5a5e-24c5-4a05-944d-d15d8b3c2b1b-kube-api-access-dwq7z\") pod \"machine-config-controller-84d6567774-glrj4\" (UID: \"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.937174 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:40 crc kubenswrapper[4809]: E1127 17:11:40.937648 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.437622253 +0000 UTC m=+136.710079605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:40 crc kubenswrapper[4809]: I1127 17:11:40.955654 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8aa01a74-34e4-4ef5-b041-17d548feb09f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6l2d\" (UID: \"8aa01a74-34e4-4ef5-b041-17d548feb09f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.039992 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.040818 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.540789029 +0000 UTC m=+136.813246381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.065902 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.073799 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.080815 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.099134 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.100338 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.119904 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.137549 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mcsp5"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.141520 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.141987 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.641969498 +0000 UTC m=+136.914426840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.146535 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.147173 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.161203 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.204873 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.212356 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.212533 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gzfwc"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.217054 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-m5lpz"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.222788 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" event={"ID":"fb3ec517-319b-4bd9-87b9-19ee21f410b6","Type":"ContainerStarted","Data":"e11d083fbf39eac27d07e2edb081c69071225d97bcdf6fc37f5fabfca6ed55df"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.226886 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" event={"ID":"6d8b54b5-a20b-4461-bca8-c8ac2ec82048","Type":"ContainerStarted","Data":"41768e20c4c7f570be531cf511bfb667eddbbaed857bc1c982df0127662fa11a"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.226933 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" event={"ID":"6d8b54b5-a20b-4461-bca8-c8ac2ec82048","Type":"ContainerStarted","Data":"3b2126a331ea39803acc287630d2117085c266c886e5f3ba0602fd0815c543c4"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.226946 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" event={"ID":"6d8b54b5-a20b-4461-bca8-c8ac2ec82048","Type":"ContainerStarted","Data":"324b5c457cd2a30cc8e5ebc8a73050f2ffe9008ea0ab018e6f6739f9e15cd3fb"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.231364 4809 generic.go:334] "Generic (PLEG): container finished" podID="54856698-dafa-4315-84c3-e0b746e815f9" containerID="b60a4bf36b8bc2af631dc7e3936596a121df78e14a8db1a2473fb3330b2b065a" exitCode=0 Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.231422 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" event={"ID":"54856698-dafa-4315-84c3-e0b746e815f9","Type":"ContainerDied","Data":"b60a4bf36b8bc2af631dc7e3936596a121df78e14a8db1a2473fb3330b2b065a"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.242953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.243340 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.743324781 +0000 UTC m=+137.015782133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.255512 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx"] Nov 27 17:11:41 crc kubenswrapper[4809]: W1127 17:11:41.265248 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86022daa_0ebf_4820_a476_e5cacc43f9d5.slice/crio-b236ee45005a0851602933bdf4fee0431436348bb398a84c8a39763ba4fba3db WatchSource:0}: Error finding container b236ee45005a0851602933bdf4fee0431436348bb398a84c8a39763ba4fba3db: Status 404 returned error can't find the container with id b236ee45005a0851602933bdf4fee0431436348bb398a84c8a39763ba4fba3db Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.268165 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" event={"ID":"dccff42e-1a6b-4239-a095-35af760b79b1","Type":"ContainerStarted","Data":"a166877d489774d5450e4d8ae30b7f47d756d3f4206c561d19e2bc4e2e620aec"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.274237 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" event={"ID":"f6349461-b5ae-4e1a-9711-9d676f1ec01f","Type":"ContainerStarted","Data":"03480f478735a3b2d35fe52cf776d0b56ed99ece8a9560c9e06ff047a4ace28e"} Nov 27 17:11:41 crc kubenswrapper[4809]: W1127 17:11:41.274726 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e615a77_dc5f_4515_8aa5_1ab5b74d6cb8.slice/crio-d1d4e1233731512e222d1e629c6e4d27828d60bc06e900dc611f6a56f939fa00 WatchSource:0}: Error finding container d1d4e1233731512e222d1e629c6e4d27828d60bc06e900dc611f6a56f939fa00: Status 404 returned error can't find the container with id d1d4e1233731512e222d1e629c6e4d27828d60bc06e900dc611f6a56f939fa00 Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.276243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" event={"ID":"23af03b4-a427-4e08-848b-0561b9cc7364","Type":"ContainerStarted","Data":"970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.276269 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" event={"ID":"23af03b4-a427-4e08-848b-0561b9cc7364","Type":"ContainerStarted","Data":"dac87afede946ab918c20020e54ac5680187e2ff813db1db1e915627022993e1"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.277496 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.280434 4809 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2cphh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.280484 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" podUID="23af03b4-a427-4e08-848b-0561b9cc7364" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.282186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" event={"ID":"2d397f9a-119c-4d88-ae8d-68ed99c919f3","Type":"ContainerStarted","Data":"96f99090ab7445504f9280bbc5e06996784e27db05442932f57aaa0d603d5ed5"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.286529 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" event={"ID":"60414207-aa6c-42ec-a02c-562a10239881","Type":"ContainerStarted","Data":"4e37d9ac74715eb115e4518d4a0fbebbcb591b59c67f8a696931720bd2aa7212"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.289334 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jtxfb" event={"ID":"fff4c3fa-a124-4cb5-806d-eea27e5dfd4a","Type":"ContainerStarted","Data":"8c8b275cb81124af690cd445900622e14cc2ca6aae43470a6ab3d249791487ca"} Nov 27 17:11:41 crc kubenswrapper[4809]: W1127 17:11:41.299893 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9a53c5a_f3a7_484e_9058_11e61da9c9ed.slice/crio-3bc9bb65e2145a64723ff1de1d4fca44b384fd27639b92fbdccaeb418bbf552f WatchSource:0}: Error finding container 3bc9bb65e2145a64723ff1de1d4fca44b384fd27639b92fbdccaeb418bbf552f: Status 404 returned error can't find the container with id 3bc9bb65e2145a64723ff1de1d4fca44b384fd27639b92fbdccaeb418bbf552f Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.314769 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" event={"ID":"dcde5369-61c9-4e6a-9d06-66a8ba0337fc","Type":"ContainerStarted","Data":"dceada8dd58645537019bdabf178018479b1b563aa4da56329d8a7524fe77ff7"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.314848 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" event={"ID":"dcde5369-61c9-4e6a-9d06-66a8ba0337fc","Type":"ContainerStarted","Data":"98ec8bc93c62724dd2c60efc0a1a9a17b77b7e2d883172b11d45fcacaf2f07c5"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.321686 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-w7f6q" event={"ID":"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7","Type":"ContainerStarted","Data":"179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.321729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-w7f6q" event={"ID":"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7","Type":"ContainerStarted","Data":"9451d2e23949e6519135565175511ad020967f65488d0e970ff5536a2fb77bec"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.328345 4809 generic.go:334] "Generic (PLEG): container finished" podID="a4dd8902-6cca-49cf-b678-6e6e0865f0dc" containerID="1e9489dfd3300304a1f0f5c1e62849e97cb649c5b526d194fc6107dc87d7cfb2" exitCode=0 Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.328560 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" event={"ID":"a4dd8902-6cca-49cf-b678-6e6e0865f0dc","Type":"ContainerDied","Data":"1e9489dfd3300304a1f0f5c1e62849e97cb649c5b526d194fc6107dc87d7cfb2"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.328709 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" event={"ID":"a4dd8902-6cca-49cf-b678-6e6e0865f0dc","Type":"ContainerStarted","Data":"2a7f2bc10bfd50e7762a2db39da54470c41f4379ddf4597343539ee39343d8d1"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.334481 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" event={"ID":"cb38e0de-3758-4e46-beb3-cf27c411725e","Type":"ContainerStarted","Data":"e10bbdad4de1050672e0a667310c1dfb0f074514eaefd5b7e50ba7177be5d8de"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.334531 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" event={"ID":"cb38e0de-3758-4e46-beb3-cf27c411725e","Type":"ContainerStarted","Data":"d7800dcc5f47643b27cbaff7031aafe192abd30f98d0182a1f926871cf8afe46"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.335642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7c7gv" event={"ID":"d94e93e8-457e-4b40-b09d-480ff253bb8a","Type":"ContainerStarted","Data":"e7a04fb9b972cbabbca567790d2b7f55efb30cfb7a71b663f3a260c9484638bf"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.341464 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zpwvm" event={"ID":"0a9bd25f-e243-4126-bddd-2a9478e35618","Type":"ContainerStarted","Data":"dc482c898b6dbabf6967495af50b209671a86679b8ef1d860e9ce90c558fa6b3"} Nov 27 17:11:41 crc kubenswrapper[4809]: W1127 17:11:41.345044 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68ba89bb_2ac5_4204_8362_891783709611.slice/crio-778f8794c67d3e72577d1d4c5458bbe953064b6c822bc9ab561b254ce8142abc WatchSource:0}: Error finding container 778f8794c67d3e72577d1d4c5458bbe953064b6c822bc9ab561b254ce8142abc: Status 404 returned error can't find the container with id 778f8794c67d3e72577d1d4c5458bbe953064b6c822bc9ab561b254ce8142abc Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.345731 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.350333 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.84658162 +0000 UTC m=+137.119038972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.350625 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" event={"ID":"08d0a292-6e77-4197-9877-72e8fae6b941","Type":"ContainerStarted","Data":"b01f3e2c3c917b31fa7639d8639a9356beeabcf3f5269bffb96733f75b3928d6"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.386139 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.401895 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" event={"ID":"be0ee2fc-9e5a-488f-8878-f29c0a7e6393","Type":"ContainerStarted","Data":"63fc9343ba615b477b8dfe69c72fe68132a9b7e09834afb11dda208d7e5489de"} Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.450400 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.453546 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:41.953529575 +0000 UTC m=+137.225986927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.455434 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.552076 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.552519 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.05249541 +0000 UTC m=+137.324952762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.565868 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-6pm2s" podStartSLOduration=117.565846275 podStartE2EDuration="1m57.565846275s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:41.556976579 +0000 UTC m=+136.829433931" watchObservedRunningTime="2025-11-27 17:11:41.565846275 +0000 UTC m=+136.838303627" Nov 27 17:11:41 crc kubenswrapper[4809]: W1127 17:11:41.608033 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb584acf2_534c_4ce1_999e_3d77468b9bd7.slice/crio-c509bfb61bfa9b6970dc5089460aea43fc8e88d8debb02a8a52aa493ba84ca4b WatchSource:0}: Error finding container c509bfb61bfa9b6970dc5089460aea43fc8e88d8debb02a8a52aa493ba84ca4b: Status 404 returned error can't find the container with id c509bfb61bfa9b6970dc5089460aea43fc8e88d8debb02a8a52aa493ba84ca4b Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.625772 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.646213 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8lfbq"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.653568 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.654048 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.154036499 +0000 UTC m=+137.426493851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.710993 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm"] Nov 27 17:11:41 crc kubenswrapper[4809]: W1127 17:11:41.715693 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29b8f288_c15b_4bfa_88fd_b7189659d57a.slice/crio-c4c9461a9e5a732812cbaffef7e2215f341e3afcf711d727f1735323573913bc WatchSource:0}: Error finding container c4c9461a9e5a732812cbaffef7e2215f341e3afcf711d727f1735323573913bc: Status 404 returned error can't find the container with id c4c9461a9e5a732812cbaffef7e2215f341e3afcf711d727f1735323573913bc Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.737408 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.747489 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6v9pw"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.759884 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.760159 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.260123729 +0000 UTC m=+137.532581081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.761212 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-w7f6q" podStartSLOduration=117.76119261 podStartE2EDuration="1m57.76119261s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:41.758121271 +0000 UTC m=+137.030578623" watchObservedRunningTime="2025-11-27 17:11:41.76119261 +0000 UTC m=+137.033649962" Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.761426 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-r7lqs"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.823388 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dpfxv"] Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.863066 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.863499 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.363483531 +0000 UTC m=+137.635940893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.964020 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.964285 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.464228257 +0000 UTC m=+137.736685609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:41 crc kubenswrapper[4809]: I1127 17:11:41.964854 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:41 crc kubenswrapper[4809]: E1127 17:11:41.965340 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.465323178 +0000 UTC m=+137.737780530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.017899 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4"] Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.070416 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.070954 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.570930755 +0000 UTC m=+137.843388107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.129838 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l4tmc" podStartSLOduration=118.129789852 podStartE2EDuration="1m58.129789852s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.077482213 +0000 UTC m=+137.349939565" watchObservedRunningTime="2025-11-27 17:11:42.129789852 +0000 UTC m=+137.402247194" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.132349 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-94nbr"] Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.166469 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-6rxcc" podStartSLOduration=117.16645305 podStartE2EDuration="1m57.16645305s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.16403611 +0000 UTC m=+137.436493462" watchObservedRunningTime="2025-11-27 17:11:42.16645305 +0000 UTC m=+137.438910402" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.170902 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x"] Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.173942 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.178579 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.678520248 +0000 UTC m=+137.950977810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.197871 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d"] Nov 27 17:11:42 crc kubenswrapper[4809]: W1127 17:11:42.275951 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92093439_c71b_4fdd_88ff_189ecd253269.slice/crio-9690739b69eb1529ffbfe37ba028e444e46a4cbe247f1f23359ca9f7f8abd171 WatchSource:0}: Error finding container 9690739b69eb1529ffbfe37ba028e444e46a4cbe247f1f23359ca9f7f8abd171: Status 404 returned error can't find the container with id 9690739b69eb1529ffbfe37ba028e444e46a4cbe247f1f23359ca9f7f8abd171 Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.279793 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.279925 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.779900492 +0000 UTC m=+138.052357844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.280122 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.280775 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.780731877 +0000 UTC m=+138.053189229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.290057 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp"] Nov 27 17:11:42 crc kubenswrapper[4809]: W1127 17:11:42.318789 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87e2c04c_53e6_4524_90b8_b122e31b9280.slice/crio-4405f68c9864fac1251034089b4f5f6df1587d78b66147125885007166d59e8f WatchSource:0}: Error finding container 4405f68c9864fac1251034089b4f5f6df1587d78b66147125885007166d59e8f: Status 404 returned error can't find the container with id 4405f68c9864fac1251034089b4f5f6df1587d78b66147125885007166d59e8f Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.349981 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd"] Nov 27 17:11:42 crc kubenswrapper[4809]: W1127 17:11:42.383985 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9e66d8f_4975_4d8c_873d_4c1e683044df.slice/crio-1683f92a92dbb0ffcd0107773a51e2a791c2320fcdc3b050dc1ebec252215685 WatchSource:0}: Error finding container 1683f92a92dbb0ffcd0107773a51e2a791c2320fcdc3b050dc1ebec252215685: Status 404 returned error can't find the container with id 1683f92a92dbb0ffcd0107773a51e2a791c2320fcdc3b050dc1ebec252215685 Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.384398 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.384856 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.884836559 +0000 UTC m=+138.157293911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: W1127 17:11:42.390430 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8aa01a74_34e4_4ef5_b041_17d548feb09f.slice/crio-e45a50f1e1fc2410a8a935b89af70e2f5b4603283680dc098a76ff29e7cd330c WatchSource:0}: Error finding container e45a50f1e1fc2410a8a935b89af70e2f5b4603283680dc098a76ff29e7cd330c: Status 404 returned error can't find the container with id e45a50f1e1fc2410a8a935b89af70e2f5b4603283680dc098a76ff29e7cd330c Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.430917 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zpwvm" event={"ID":"0a9bd25f-e243-4126-bddd-2a9478e35618","Type":"ContainerStarted","Data":"09a4c7104ae35050bd32f5e119afc4ad98bbf419cd979cbd3aaef248991a4314"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.448473 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" event={"ID":"b584acf2-534c-4ce1-999e-3d77468b9bd7","Type":"ContainerStarted","Data":"842368af0aac26084acd63250c72d0a5c1146745c8c1d7c109b1787a79439737"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.448523 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" event={"ID":"b584acf2-534c-4ce1-999e-3d77468b9bd7","Type":"ContainerStarted","Data":"c509bfb61bfa9b6970dc5089460aea43fc8e88d8debb02a8a52aa493ba84ca4b"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.457280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" event={"ID":"f6349461-b5ae-4e1a-9711-9d676f1ec01f","Type":"ContainerStarted","Data":"f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.459018 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.462642 4809 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-wq2f9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.463567 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" podUID="f6349461-b5ae-4e1a-9711-9d676f1ec01f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.468623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" event={"ID":"f62405dd-1836-4b7d-a785-27f34188a241","Type":"ContainerStarted","Data":"1d1711cf8f9f63043779cdb49d93d62651ff85feb50108d05179f2e55b9d8457"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.478517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jtxfb" event={"ID":"fff4c3fa-a124-4cb5-806d-eea27e5dfd4a","Type":"ContainerStarted","Data":"43af0d62b2cecea915e22251a7bf71229c01656d130272f290abe0681e90c3cc"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.482954 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-jtxfb" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.485613 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-jtxfb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.485646 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jtxfb" podUID="fff4c3fa-a124-4cb5-806d-eea27e5dfd4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.486499 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.486884 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:42.986873702 +0000 UTC m=+138.259331054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.489211 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" event={"ID":"1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8","Type":"ContainerStarted","Data":"d1d4e1233731512e222d1e629c6e4d27828d60bc06e900dc611f6a56f939fa00"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.498227 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" event={"ID":"8aa01a74-34e4-4ef5-b041-17d548feb09f","Type":"ContainerStarted","Data":"e45a50f1e1fc2410a8a935b89af70e2f5b4603283680dc098a76ff29e7cd330c"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.502089 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" event={"ID":"08d0a292-6e77-4197-9877-72e8fae6b941","Type":"ContainerStarted","Data":"6126a15287c5713af5342da382ba71f664cc0ca3d2f1be2a1a640a2a3188265b"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.511180 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" event={"ID":"fb3ec517-319b-4bd9-87b9-19ee21f410b6","Type":"ContainerStarted","Data":"78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.512163 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.513492 4809 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2t2fb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.513537 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" podUID="fb3ec517-319b-4bd9-87b9-19ee21f410b6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.514435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" event={"ID":"92093439-c71b-4fdd-88ff-189ecd253269","Type":"ContainerStarted","Data":"9690739b69eb1529ffbfe37ba028e444e46a4cbe247f1f23359ca9f7f8abd171"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.542907 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" event={"ID":"d9a53c5a-f3a7-484e-9058-11e61da9c9ed","Type":"ContainerStarted","Data":"c580c4130bf11ccca5d7c3e2a5894d779104ac3e817def37324f423f7d3299a6"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.542971 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" event={"ID":"d9a53c5a-f3a7-484e-9058-11e61da9c9ed","Type":"ContainerStarted","Data":"3bc9bb65e2145a64723ff1de1d4fca44b384fd27639b92fbdccaeb418bbf552f"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.546668 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" event={"ID":"1793bd02-489b-422c-9541-4354b4c4c020","Type":"ContainerStarted","Data":"e394528278430fe815840736befce3f7a0d9a20e118b86f649b27ab1c215d543"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.549401 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" event={"ID":"4f877f18-1c0c-43b7-9873-9bd3d48012d4","Type":"ContainerStarted","Data":"8ac7746eb78f6f61ecefc5f81e1e1943cdf94a69edea1ec43d8a93fb238cff3f"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.557627 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" event={"ID":"be0ee2fc-9e5a-488f-8878-f29c0a7e6393","Type":"ContainerStarted","Data":"ebecd71925329ed74cda0e6ddf1cedd27960ccb731312c632f6b3891c49cd2ba"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.587557 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.589628 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.089607416 +0000 UTC m=+138.362064768 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.599965 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6v9pw" event={"ID":"f6628245-a537-4cf1-8b4c-207bb7cb3d0b","Type":"ContainerStarted","Data":"46aa77f1b12f7b1634f6cd64250504d061b88da0ddbf748915efcff9ae476631"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.611176 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" podStartSLOduration=118.611141047 podStartE2EDuration="1m58.611141047s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.610847449 +0000 UTC m=+137.883304801" watchObservedRunningTime="2025-11-27 17:11:42.611141047 +0000 UTC m=+137.883598419" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.612183 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" podStartSLOduration=117.612173797 podStartE2EDuration="1m57.612173797s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.557977274 +0000 UTC m=+137.830434656" watchObservedRunningTime="2025-11-27 17:11:42.612173797 +0000 UTC m=+137.884631159" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.615455 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" event={"ID":"02f0a77f-7099-480e-b4e6-6183592ca5e6","Type":"ContainerStarted","Data":"1ba160c4f5aff6b636a61ec6edb6b00a20c6660cb2093ca884a0d5b7a1ac5cb6"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.623418 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" event={"ID":"b9e66d8f-4975-4d8c-873d-4c1e683044df","Type":"ContainerStarted","Data":"1683f92a92dbb0ffcd0107773a51e2a791c2320fcdc3b050dc1ebec252215685"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.625187 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" event={"ID":"29b8f288-c15b-4bfa-88fd-b7189659d57a","Type":"ContainerStarted","Data":"c4c9461a9e5a732812cbaffef7e2215f341e3afcf711d727f1735323573913bc"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.626797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7c7gv" event={"ID":"d94e93e8-457e-4b40-b09d-480ff253bb8a","Type":"ContainerStarted","Data":"517100ba97b02b836ba65666a26314fdabdd6ea4016e25aa43eba5075251fd92"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.627560 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.630860 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" event={"ID":"2d397f9a-119c-4d88-ae8d-68ed99c919f3","Type":"ContainerStarted","Data":"16c1a12188a7a8fc435c741ec492355291d131c9a2c4635aa1778fadf819145b"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.634431 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-7c7gv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.634474 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7c7gv" podUID="d94e93e8-457e-4b40-b09d-480ff253bb8a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.644687 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" event={"ID":"d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4","Type":"ContainerStarted","Data":"6b8e16d87209e4f7c8d0229155311f17387f6f4a4bee83038c362f0bc681e074"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.644781 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" event={"ID":"d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4","Type":"ContainerStarted","Data":"93fe10b3e048986520bdc41320b81fcc880db94ccb5fbd8c9cb1835b4ef18cb6"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.645840 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-jtxfb" podStartSLOduration=118.645827918 podStartE2EDuration="1m58.645827918s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.640438062 +0000 UTC m=+137.912895414" watchObservedRunningTime="2025-11-27 17:11:42.645827918 +0000 UTC m=+137.918285270" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.670568 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" event={"ID":"dccff42e-1a6b-4239-a095-35af760b79b1","Type":"ContainerStarted","Data":"9e92f16f7d9e4eac1268885b49057b08d320e60536fa4529f3703916403f580a"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.677396 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xbsxf" podStartSLOduration=117.677368807 podStartE2EDuration="1m57.677368807s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.675325839 +0000 UTC m=+137.947783191" watchObservedRunningTime="2025-11-27 17:11:42.677368807 +0000 UTC m=+137.949826159" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.678258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" event={"ID":"93429dc0-c4aa-4eec-a48d-7128e13d0d8f","Type":"ContainerStarted","Data":"046940c1889048c01f2eceef6e9087370fd9ba43a8a6a262914232186bda603a"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.688997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" event={"ID":"86022daa-0ebf-4820-a476-e5cacc43f9d5","Type":"ContainerStarted","Data":"915bf726b3678b5fcc7a7216e67fb8090079d341eee9b483b0372dcf93f73249"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.689073 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" event={"ID":"86022daa-0ebf-4820-a476-e5cacc43f9d5","Type":"ContainerStarted","Data":"b236ee45005a0851602933bdf4fee0431436348bb398a84c8a39763ba4fba3db"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.690272 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.693419 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.19340162 +0000 UTC m=+138.465858962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.702964 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" event={"ID":"68ba89bb-2ac5-4204-8362-891783709611","Type":"ContainerStarted","Data":"778f8794c67d3e72577d1d4c5458bbe953064b6c822bc9ab561b254ce8142abc"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.717292 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kzdxz" event={"ID":"9a3cce9e-247e-457d-b41c-a1a7ee3c7888","Type":"ContainerStarted","Data":"361bd6aa1c61895dc57f48dd61801c086007ef5de0e16db9e541ef5bc2bf5532"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.717343 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kzdxz" event={"ID":"9a3cce9e-247e-457d-b41c-a1a7ee3c7888","Type":"ContainerStarted","Data":"a8ce2deacfd96ec380cb49b19dd5f999bf4a62d7be512083bcdba54e6dc50c5c"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.722499 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" podStartSLOduration=117.722484158 podStartE2EDuration="1m57.722484158s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.722308233 +0000 UTC m=+137.994765585" watchObservedRunningTime="2025-11-27 17:11:42.722484158 +0000 UTC m=+137.994941510" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.733026 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" event={"ID":"cbb30a57-2b12-464b-b9aa-6a54d8314abb","Type":"ContainerStarted","Data":"91ce961e2d12e7d255955e91eb7d027d90110f9ce28c5498a67a0bf105460fda"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.735530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" event={"ID":"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b","Type":"ContainerStarted","Data":"1dca7def431e42b91688f9a88aef562bede8f2182b14cd7fe4f14d3b97a762cd"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.739287 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-r7lqs" event={"ID":"73315a03-8ebc-44c1-bc28-0fa0965b1753","Type":"ContainerStarted","Data":"96dfdb3254cfc9b4825a10ec977427186fe6b275aaf09470da17019dba7a13e8"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.746565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" event={"ID":"87e2c04c-53e6-4524-90b8-b122e31b9280","Type":"ContainerStarted","Data":"4405f68c9864fac1251034089b4f5f6df1587d78b66147125885007166d59e8f"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.760164 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" event={"ID":"60414207-aa6c-42ec-a02c-562a10239881","Type":"ContainerStarted","Data":"45432d041cc6722850b0ed82be007fd8668b5ebf1838a19cdd7b44ef05e0252e"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.766069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" event={"ID":"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5","Type":"ContainerStarted","Data":"116e81aa3c967d0587a14729e408ff5ebe09e9be691652c397375624122cd5e2"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.792204 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.792358 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.292322333 +0000 UTC m=+138.564779675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.792440 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.793147 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.293127767 +0000 UTC m=+138.565585109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.804626 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p" event={"ID":"132b91ef-f529-4194-b6cd-f356fc4d0e33","Type":"ContainerStarted","Data":"0b0c48ae4f8e5f192a6de800c386019cea81d9b4ba10070e72c0ec602e091af6"} Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.825461 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.844355 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-7c7gv" podStartSLOduration=118.844331384 podStartE2EDuration="1m58.844331384s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.842702856 +0000 UTC m=+138.115160208" watchObservedRunningTime="2025-11-27 17:11:42.844331384 +0000 UTC m=+138.116788736" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.850295 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.850834 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mdsb" podStartSLOduration=117.85080628 podStartE2EDuration="1m57.85080628s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.808090298 +0000 UTC m=+138.080547650" watchObservedRunningTime="2025-11-27 17:11:42.85080628 +0000 UTC m=+138.123263632" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.856525 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.856591 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.895084 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.895262 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.395232042 +0000 UTC m=+138.667689394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.895874 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:42 crc kubenswrapper[4809]: E1127 17:11:42.897323 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.397315122 +0000 UTC m=+138.669772474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.937671 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" podStartSLOduration=118.937645965 podStartE2EDuration="1m58.937645965s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.92460477 +0000 UTC m=+138.197062122" watchObservedRunningTime="2025-11-27 17:11:42.937645965 +0000 UTC m=+138.210103317" Nov 27 17:11:42 crc kubenswrapper[4809]: I1127 17:11:42.939409 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzk7p" podStartSLOduration=117.939398556 podStartE2EDuration="1m57.939398556s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.891573417 +0000 UTC m=+138.164030779" watchObservedRunningTime="2025-11-27 17:11:42.939398556 +0000 UTC m=+138.211855908" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:42.999200 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:42.999326 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.499302154 +0000 UTC m=+138.771759506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.000203 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.000609 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.500588921 +0000 UTC m=+138.773046283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.073655 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kzwwp" podStartSLOduration=118.073637838 podStartE2EDuration="1m58.073637838s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:43.072820515 +0000 UTC m=+138.345277867" watchObservedRunningTime="2025-11-27 17:11:43.073637838 +0000 UTC m=+138.346095190" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.074356 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-kzdxz" podStartSLOduration=118.074350229 podStartE2EDuration="1m58.074350229s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:42.981647005 +0000 UTC m=+138.254104357" watchObservedRunningTime="2025-11-27 17:11:43.074350229 +0000 UTC m=+138.346807581" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.101427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.101720 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.601702718 +0000 UTC m=+138.874160060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.101764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.102098 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.602091799 +0000 UTC m=+138.874549151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.207269 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.208102 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.708087717 +0000 UTC m=+138.980545059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.315647 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.316326 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.816293057 +0000 UTC m=+139.088750409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.416790 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.417240 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:43.917223229 +0000 UTC m=+139.189680581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.518860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.519880 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.01986733 +0000 UTC m=+139.292324682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.620548 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.621016 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.121001087 +0000 UTC m=+139.393458429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.721815 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.722160 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.222145034 +0000 UTC m=+139.494602386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.824930 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.825175 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.325160896 +0000 UTC m=+139.597618248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.825218 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.825623 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.325591419 +0000 UTC m=+139.598048771 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.832152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" event={"ID":"02f0a77f-7099-480e-b4e6-6183592ca5e6","Type":"ContainerStarted","Data":"795851424a678bdc27022fda09dbe48431db424416d37d04d9e37c4ee10a1cbc"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.836207 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" event={"ID":"92093439-c71b-4fdd-88ff-189ecd253269","Type":"ContainerStarted","Data":"25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.837144 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.842900 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-94nbr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.842962 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" podUID="92093439-c71b-4fdd-88ff-189ecd253269" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.853432 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" event={"ID":"87e2c04c-53e6-4524-90b8-b122e31b9280","Type":"ContainerStarted","Data":"ebf4b8bbcc00ee2fbe768d2c525dd2f15c575059caf55a953bd8135fbbcb605c"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.854596 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.854630 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.854965 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-m5lpz" podStartSLOduration=118.854950886 podStartE2EDuration="1m58.854950886s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:43.854332717 +0000 UTC m=+139.126790069" watchObservedRunningTime="2025-11-27 17:11:43.854950886 +0000 UTC m=+139.127408258" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.868010 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p" event={"ID":"132b91ef-f529-4194-b6cd-f356fc4d0e33","Type":"ContainerStarted","Data":"4d6f5bae6c4ae8f00e42a0e958791c6d54e911a20e37d9a3586ece18d44bbc81"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.875382 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" event={"ID":"1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8","Type":"ContainerStarted","Data":"984f251b32e784d1fbf8dfae5e8b8bdb098654ff9787f04fece9531d8c5535ed"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.879190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" event={"ID":"15c3167d-77d3-457f-b602-5e87f96afe8d","Type":"ContainerStarted","Data":"3c3714c1a6319450a3525d33ff2a1a82e6ca523ba356cf684a9d85b33b861836"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.879243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" event={"ID":"15c3167d-77d3-457f-b602-5e87f96afe8d","Type":"ContainerStarted","Data":"259e9d4bbad6c573d5ccac10badb7f2c191140864b5825fcef72c719c50f449d"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.880490 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6v9pw" event={"ID":"f6628245-a537-4cf1-8b4c-207bb7cb3d0b","Type":"ContainerStarted","Data":"a1fd9010a50cfd08b5d10729835083ab66a7767b1a922d618ac355188553fa1b"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.882654 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tvl4x" podStartSLOduration=118.882644995 podStartE2EDuration="1m58.882644995s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:43.880946365 +0000 UTC m=+139.153403707" watchObservedRunningTime="2025-11-27 17:11:43.882644995 +0000 UTC m=+139.155102337" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.898545 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" event={"ID":"1793bd02-489b-422c-9541-4354b4c4c020","Type":"ContainerStarted","Data":"e477e4a402b962c3abe3412936ab3afcb3615ee7931b24b5089cd01a5160f125"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.899166 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.909312 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" event={"ID":"f62405dd-1836-4b7d-a785-27f34188a241","Type":"ContainerStarted","Data":"9c8d99a7cd67b73a08783eaf1421abfd4e4fb852e2784f285f799bbad040b9b5"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.913839 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" event={"ID":"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b","Type":"ContainerStarted","Data":"38e0f0d4b791e2445acda4132fcec3cd8431a03d4797a41ad46c7331062467e3"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.913818 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" podStartSLOduration=118.913801784 podStartE2EDuration="1m58.913801784s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:43.913466824 +0000 UTC m=+139.185924176" watchObservedRunningTime="2025-11-27 17:11:43.913801784 +0000 UTC m=+139.186259136" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.915189 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-j6ggx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.915279 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" podUID="1793bd02-489b-422c-9541-4354b4c4c020" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.916816 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" event={"ID":"a4dd8902-6cca-49cf-b678-6e6e0865f0dc","Type":"ContainerStarted","Data":"48ee780b6a4ac8518552c43d465b435cb5c5ea8f36a33c8d844d4ff342fbb04c"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.926963 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.927092 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.427068015 +0000 UTC m=+139.699525367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.927423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.928867 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" event={"ID":"93429dc0-c4aa-4eec-a48d-7128e13d0d8f","Type":"ContainerStarted","Data":"97f665c345ba47f58aff8f9dbde3385552b3dea75206f8dc7cb39ff45f9f7aca"} Nov 27 17:11:43 crc kubenswrapper[4809]: E1127 17:11:43.929395 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.429375922 +0000 UTC m=+139.701833274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.934655 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" event={"ID":"cbb30a57-2b12-464b-b9aa-6a54d8314abb","Type":"ContainerStarted","Data":"c048d9915e5c6e579c482f4cf16de0e1c6003ae0bebb528244ca0f24fb7b1cdd"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.935768 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.940491 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zvlsm" podStartSLOduration=118.940470542 podStartE2EDuration="1m58.940470542s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:43.936254721 +0000 UTC m=+139.208712083" watchObservedRunningTime="2025-11-27 17:11:43.940470542 +0000 UTC m=+139.212927904" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.941154 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-qt7fm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.941208 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" podUID="cbb30a57-2b12-464b-b9aa-6a54d8314abb" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.945374 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-r7lqs" event={"ID":"73315a03-8ebc-44c1-bc28-0fa0965b1753","Type":"ContainerStarted","Data":"e79b81142c73af6f9351976187a7343cfa3e622e64959f201d4e841fd8760cfc"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.949669 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" event={"ID":"4f877f18-1c0c-43b7-9873-9bd3d48012d4","Type":"ContainerStarted","Data":"956f0cf1e42806840bba8220d91f6f7878b10e9bbf766f1132fa8132a8983886"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.951149 4809 generic.go:334] "Generic (PLEG): container finished" podID="d9a53c5a-f3a7-484e-9058-11e61da9c9ed" containerID="c580c4130bf11ccca5d7c3e2a5894d779104ac3e817def37324f423f7d3299a6" exitCode=0 Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.951216 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" event={"ID":"d9a53c5a-f3a7-484e-9058-11e61da9c9ed","Type":"ContainerDied","Data":"c580c4130bf11ccca5d7c3e2a5894d779104ac3e817def37324f423f7d3299a6"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.953305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" event={"ID":"b9e66d8f-4975-4d8c-873d-4c1e683044df","Type":"ContainerStarted","Data":"05ef0b10a6d65a01cda9af4e6d8cc05aed421985fa4fa5a800a6e65bb2d6e45c"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.957322 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" podStartSLOduration=118.957302908 podStartE2EDuration="1m58.957302908s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:43.957258617 +0000 UTC m=+139.229715979" watchObservedRunningTime="2025-11-27 17:11:43.957302908 +0000 UTC m=+139.229760250" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.960008 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hxvj9" event={"ID":"68ba89bb-2ac5-4204-8362-891783709611","Type":"ContainerStarted","Data":"f5bfa94f9fb073645cd439fa076e180262f8e696eba6981c1b58055961ec30c9"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.966632 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" event={"ID":"54856698-dafa-4315-84c3-e0b746e815f9","Type":"ContainerStarted","Data":"b4432774dd778fe734d052fa47efabeff1c6c5edf756ff768766da59b2c6fd27"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.968240 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" event={"ID":"29b8f288-c15b-4bfa-88fd-b7189659d57a","Type":"ContainerStarted","Data":"67811d4e72f116807e6a13fb4cd9f81ae8d95c83231b677de6db16e65b10ab3f"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.969970 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" event={"ID":"be0ee2fc-9e5a-488f-8878-f29c0a7e6393","Type":"ContainerStarted","Data":"b29dbc6b87271c5799dd8d69c944a5ed22b12b4f143f10f616debedb64c64175"} Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.970683 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-jtxfb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.970731 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jtxfb" podUID="fff4c3fa-a124-4cb5-806d-eea27e5dfd4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.970914 4809 patch_prober.go:28] interesting pod/console-operator-58897d9998-7c7gv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.970988 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7c7gv" podUID="d94e93e8-457e-4b40-b09d-480ff253bb8a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.971356 4809 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2t2fb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.971424 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" podUID="fb3ec517-319b-4bd9-87b9-19ee21f410b6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Nov 27 17:11:43 crc kubenswrapper[4809]: I1127 17:11:43.996341 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kwxf5" podStartSLOduration=118.996311283 podStartE2EDuration="1m58.996311283s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:43.994373227 +0000 UTC m=+139.266830569" watchObservedRunningTime="2025-11-27 17:11:43.996311283 +0000 UTC m=+139.268768635" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.029712 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.030273 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.530236982 +0000 UTC m=+139.802694334 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.030941 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.033271 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.533255999 +0000 UTC m=+139.805713341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.033368 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.043710 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-r7lqs" podStartSLOduration=7.043668509 podStartE2EDuration="7.043668509s" podCreationTimestamp="2025-11-27 17:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:44.022181819 +0000 UTC m=+139.294639171" watchObservedRunningTime="2025-11-27 17:11:44.043668509 +0000 UTC m=+139.316125861" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.045324 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" podStartSLOduration=119.045317437 podStartE2EDuration="1m59.045317437s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:44.041859027 +0000 UTC m=+139.314316379" watchObservedRunningTime="2025-11-27 17:11:44.045317437 +0000 UTC m=+139.317774789" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.072597 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" podStartSLOduration=119.072569793 podStartE2EDuration="1m59.072569793s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:44.068465235 +0000 UTC m=+139.340922597" watchObservedRunningTime="2025-11-27 17:11:44.072569793 +0000 UTC m=+139.345027145" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.111038 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ppfvt" podStartSLOduration=119.111011302 podStartE2EDuration="1m59.111011302s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:44.104333959 +0000 UTC m=+139.376791311" watchObservedRunningTime="2025-11-27 17:11:44.111011302 +0000 UTC m=+139.383468654" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.132505 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.132758 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.632713758 +0000 UTC m=+139.905171110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.133134 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.136694 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.636673602 +0000 UTC m=+139.909130954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.139288 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" podStartSLOduration=119.139255956 podStartE2EDuration="1m59.139255956s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:44.136888718 +0000 UTC m=+139.409346070" watchObservedRunningTime="2025-11-27 17:11:44.139255956 +0000 UTC m=+139.411713308" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.159225 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-zpwvm" podStartSLOduration=7.159180571 podStartE2EDuration="7.159180571s" podCreationTimestamp="2025-11-27 17:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:44.157537024 +0000 UTC m=+139.429994376" watchObservedRunningTime="2025-11-27 17:11:44.159180571 +0000 UTC m=+139.431637923" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.187407 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-8lfbq" podStartSLOduration=119.187382214 podStartE2EDuration="1m59.187382214s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:44.186321134 +0000 UTC m=+139.458778486" watchObservedRunningTime="2025-11-27 17:11:44.187382214 +0000 UTC m=+139.459839566" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.234626 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.234976 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.734943806 +0000 UTC m=+140.007401158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.336204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.336566 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.836551488 +0000 UTC m=+140.109008840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.438782 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.439835 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:44.939807756 +0000 UTC m=+140.212265108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.541234 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.541760 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.041727786 +0000 UTC m=+140.314185138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.643237 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.644210 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.144193081 +0000 UTC m=+140.416650433 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.745909 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.746429 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.2464096 +0000 UTC m=+140.518866952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.815068 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.815167 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.817790 4809 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-wpq4q container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.817858 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" podUID="54856698-dafa-4315-84c3-e0b746e815f9" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.847986 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.848247 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.348205936 +0000 UTC m=+140.620663288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.848583 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.849099 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.349090563 +0000 UTC m=+140.621547915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.851314 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.851366 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.950180 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.950300 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.450271811 +0000 UTC m=+140.722729153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.951544 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:44 crc kubenswrapper[4809]: E1127 17:11:44.951988 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.4519795 +0000 UTC m=+140.724436852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.976615 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p" event={"ID":"132b91ef-f529-4194-b6cd-f356fc4d0e33","Type":"ContainerStarted","Data":"71e1747bba798a9a63ab4f61322ad9a005e8c37a77e9dc3d639d8016dc40d0e5"} Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.978841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" event={"ID":"4f877f18-1c0c-43b7-9873-9bd3d48012d4","Type":"ContainerStarted","Data":"4e3ab0a3c725543f7fc6815be52b996da8fe812c81b4393eb96cefa3cd690480"} Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.979777 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.981402 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6v9pw" event={"ID":"f6628245-a537-4cf1-8b4c-207bb7cb3d0b","Type":"ContainerStarted","Data":"ab8dafb48e12885e349f0681d1b56bdc44cd0a4c13776bd98baf3b389e4ab59c"} Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.982153 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.983580 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" event={"ID":"455e5a5e-24c5-4a05-944d-d15d8b3c2b1b","Type":"ContainerStarted","Data":"fbb88f0d508623804ff64fabce59747b8c9f41a90b107a95e224b3942505ddab"} Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.986186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" event={"ID":"86022daa-0ebf-4820-a476-e5cacc43f9d5","Type":"ContainerStarted","Data":"1e332dc8c6c4f727c5857d87f2b282b54f5178875572c718bcb1627f167076cf"} Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.988950 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" event={"ID":"a4dd8902-6cca-49cf-b678-6e6e0865f0dc","Type":"ContainerStarted","Data":"8d6284f0d5b88fe4be7428ce05efa04ca9a5c7e96d0b48da8f346f7088c73f2c"} Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.991527 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" event={"ID":"b9e66d8f-4975-4d8c-873d-4c1e683044df","Type":"ContainerStarted","Data":"24136639133d0cfb309a9cc905754c8f2e93e2e687141fa7283f9cbc4fff3cc6"} Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.993664 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" event={"ID":"8aa01a74-34e4-4ef5-b041-17d548feb09f","Type":"ContainerStarted","Data":"37801369e4bef64cec954da285c937c4bb767f47957a2601072b4d99ae236bb2"} Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.996039 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" event={"ID":"d9a53c5a-f3a7-484e-9058-11e61da9c9ed","Type":"ContainerStarted","Data":"501d45be2a9fadfd7e0b78625241b8dbfb59486df19c9d9e8fe39ca39d1457f0"} Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.996425 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:44 crc kubenswrapper[4809]: I1127 17:11:44.998715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" event={"ID":"d6c1b4bb-6eb1-4951-9dd6-ab2d4d6231f4","Type":"ContainerStarted","Data":"b705989e3cb97092166e3311beddf4db73175db9e1955404e371737e6c5de737"} Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.002203 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-j6ggx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.002280 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" podUID="1793bd02-489b-422c-9541-4354b4c4c020" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.002600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" event={"ID":"1e615a77-dc5f-4515-8aa5-1ab5b74d6cb8","Type":"ContainerStarted","Data":"e57b44ab5c8bd1e69fa5146da445d9b628101e52d0ab363764366f238a84be37"} Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.003245 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-jtxfb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.003299 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jtxfb" podUID="fff4c3fa-a124-4cb5-806d-eea27e5dfd4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.003253 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-qt7fm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.003359 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" podUID="cbb30a57-2b12-464b-b9aa-6a54d8314abb" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.003433 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-94nbr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.003489 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" podUID="92093439-c71b-4fdd-88ff-189ecd253269" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.054056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.054305 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.55425171 +0000 UTC m=+140.826709062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.054842 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.059580 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.559555604 +0000 UTC m=+140.832012966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.092242 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.113069 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tf68p" podStartSLOduration=120.113047836 podStartE2EDuration="2m0.113047836s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.07192496 +0000 UTC m=+140.344382312" watchObservedRunningTime="2025-11-27 17:11:45.113047836 +0000 UTC m=+140.385505208" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.150235 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-gzfwc" podStartSLOduration=120.150211208 podStartE2EDuration="2m0.150211208s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.114007154 +0000 UTC m=+140.386464516" watchObservedRunningTime="2025-11-27 17:11:45.150211208 +0000 UTC m=+140.422668550" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.157214 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.158729 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.658705143 +0000 UTC m=+140.931162495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.162042 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-glrj4" podStartSLOduration=120.162009738 podStartE2EDuration="2m0.162009738s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.143645049 +0000 UTC m=+140.416102401" watchObservedRunningTime="2025-11-27 17:11:45.162009738 +0000 UTC m=+140.434467090" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.187444 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" podStartSLOduration=120.187420731 podStartE2EDuration="2m0.187420731s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.183049166 +0000 UTC m=+140.455506518" watchObservedRunningTime="2025-11-27 17:11:45.187420731 +0000 UTC m=+140.459878083" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.221634 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6l2d" podStartSLOduration=120.221614638 podStartE2EDuration="2m0.221614638s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.218809217 +0000 UTC m=+140.491266569" watchObservedRunningTime="2025-11-27 17:11:45.221614638 +0000 UTC m=+140.494072000" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.258394 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" podStartSLOduration=121.258378789 podStartE2EDuration="2m1.258378789s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.256637648 +0000 UTC m=+140.529095000" watchObservedRunningTime="2025-11-27 17:11:45.258378789 +0000 UTC m=+140.530836131" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.260214 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.260581 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.760566151 +0000 UTC m=+141.033023503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.281838 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-mcsp5" podStartSLOduration=120.281816534 podStartE2EDuration="2m0.281816534s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.276267554 +0000 UTC m=+140.548724906" watchObservedRunningTime="2025-11-27 17:11:45.281816534 +0000 UTC m=+140.554273886" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.313843 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ngdqp" podStartSLOduration=120.313808267 podStartE2EDuration="2m0.313808267s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.311200103 +0000 UTC m=+140.583657455" watchObservedRunningTime="2025-11-27 17:11:45.313808267 +0000 UTC m=+140.586265619" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.361830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.362334 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.862318777 +0000 UTC m=+141.134776119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.369950 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s4lpp" podStartSLOduration=121.369931327 podStartE2EDuration="2m1.369931327s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.339967602 +0000 UTC m=+140.612424954" watchObservedRunningTime="2025-11-27 17:11:45.369931327 +0000 UTC m=+140.642388679" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.394606 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6v9pw" podStartSLOduration=8.394587007 podStartE2EDuration="8.394587007s" podCreationTimestamp="2025-11-27 17:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.370467312 +0000 UTC m=+140.642924664" watchObservedRunningTime="2025-11-27 17:11:45.394587007 +0000 UTC m=+140.667044359" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.429554 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" podStartSLOduration=120.429537986 podStartE2EDuration="2m0.429537986s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.42797029 +0000 UTC m=+140.700427642" watchObservedRunningTime="2025-11-27 17:11:45.429537986 +0000 UTC m=+140.701995338" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.430727 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" podStartSLOduration=121.43072152 podStartE2EDuration="2m1.43072152s" podCreationTimestamp="2025-11-27 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:45.395649968 +0000 UTC m=+140.668107340" watchObservedRunningTime="2025-11-27 17:11:45.43072152 +0000 UTC m=+140.703178872" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.463935 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.464701 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:45.964688309 +0000 UTC m=+141.237145661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.565295 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.565517 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.065483557 +0000 UTC m=+141.337940909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.566032 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.566412 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.066395124 +0000 UTC m=+141.338852476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.667230 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.667484 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.167416698 +0000 UTC m=+141.439874060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.667705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.668681 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.168670003 +0000 UTC m=+141.441127365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.769193 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.769411 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.269376968 +0000 UTC m=+141.541834320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.769568 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.769876 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.269864752 +0000 UTC m=+141.542322104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.856322 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 17:11:45 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 27 17:11:45 crc kubenswrapper[4809]: [+]process-running ok Nov 27 17:11:45 crc kubenswrapper[4809]: healthz check failed Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.857109 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.871319 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.871554 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.371514415 +0000 UTC m=+141.643971767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:45 crc kubenswrapper[4809]: I1127 17:11:45.897872 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:45 crc kubenswrapper[4809]: E1127 17:11:45.898764 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.398730159 +0000 UTC m=+141.671187511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.002556 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.003228 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.503203464 +0000 UTC m=+141.775660816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.022410 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" event={"ID":"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5","Type":"ContainerStarted","Data":"467a290ec43542044a1e0e121004280dcbaa92e59907b76a613a8d6cd01a3605"} Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.023464 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-qt7fm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.023527 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" podUID="cbb30a57-2b12-464b-b9aa-6a54d8314abb" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.026375 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-94nbr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.026453 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" podUID="92093439-c71b-4fdd-88ff-189ecd253269" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.106126 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.115955 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.615883904 +0000 UTC m=+141.888341256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.210197 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.210507 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.710487653 +0000 UTC m=+141.982945005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.210892 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.211325 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.711317887 +0000 UTC m=+141.983775239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.311375 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.311646 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.81162987 +0000 UTC m=+142.084087232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.412649 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.413189 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:46.913166639 +0000 UTC m=+142.185623991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.503149 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.504386 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.510677 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.510695 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.515958 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.516121 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.016093978 +0000 UTC m=+142.288551330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.516205 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.516303 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.516335 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.516704 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.016692665 +0000 UTC m=+142.289150197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.522616 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.617910 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.618138 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.11810306 +0000 UTC m=+142.390560412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.618476 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.618517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.618599 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.618611 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.619000 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.118991466 +0000 UTC m=+142.391448818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.693934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.719790 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.719996 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.219952698 +0000 UTC m=+142.492410050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.720632 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.720963 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.220949617 +0000 UTC m=+142.493406969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.821149 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.821416 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.321375104 +0000 UTC m=+142.593832456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.821676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.822128 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.322110435 +0000 UTC m=+142.594567857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.835387 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.859161 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 17:11:46 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 27 17:11:46 crc kubenswrapper[4809]: [+]process-running ok Nov 27 17:11:46 crc kubenswrapper[4809]: healthz check failed Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.859220 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.922814 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.923036 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.422993885 +0000 UTC m=+142.695451237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:46 crc kubenswrapper[4809]: I1127 17:11:46.923199 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:46 crc kubenswrapper[4809]: E1127 17:11:46.923572 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.423564632 +0000 UTC m=+142.696021984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.024453 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.024641 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.524616607 +0000 UTC m=+142.797073959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.024816 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.025200 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.525191783 +0000 UTC m=+142.797649135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.045397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" event={"ID":"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5","Type":"ContainerStarted","Data":"6edeed8bf047ac4bfa59bd8fdf777048f4f19d94bc17201e1559e350989ab8b9"} Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.133436 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.133762 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.633708223 +0000 UTC m=+142.906165575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.134239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.136524 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.636512164 +0000 UTC m=+142.908969516 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.236339 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.236692 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.736627552 +0000 UTC m=+143.009084904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.237107 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.237761 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.737733904 +0000 UTC m=+143.010191246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.319042 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.338971 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.339509 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.83948522 +0000 UTC m=+143.111942562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.441759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.442172 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:47.94213715 +0000 UTC m=+143.214594572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.545817 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.546140 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.046089889 +0000 UTC m=+143.318547241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.648101 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.648557 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.148537224 +0000 UTC m=+143.420994576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.749446 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.749665 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.2496314 +0000 UTC m=+143.522088752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.750024 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.750403 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.250394902 +0000 UTC m=+143.522852254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.851883 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.852409 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.352383904 +0000 UTC m=+143.624841256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.852823 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hrw24"] Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.853937 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.855291 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 17:11:47 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 27 17:11:47 crc kubenswrapper[4809]: [+]process-running ok Nov 27 17:11:47 crc kubenswrapper[4809]: healthz check failed Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.855336 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 17:11:47 crc kubenswrapper[4809]: W1127 17:11:47.865121 4809 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.865174 4809 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.953943 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-utilities\") pod \"community-operators-hrw24\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.954012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.954040 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8g96\" (UniqueName: \"kubernetes.io/projected/b1d9678d-171a-427b-9b37-192f523e63b2-kube-api-access-x8g96\") pod \"community-operators-hrw24\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.954107 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-catalog-content\") pod \"community-operators-hrw24\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:47 crc kubenswrapper[4809]: E1127 17:11:47.954487 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.454463109 +0000 UTC m=+143.726920461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:47 crc kubenswrapper[4809]: I1127 17:11:47.988948 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrw24"] Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.051316 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" event={"ID":"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5","Type":"ContainerStarted","Data":"b88c6c7768bd625c70d2ce5339a0c4226a32d92e2270670bfd35a459a8ab83db"} Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.053005 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ee4af418-0d2a-4797-9e6d-00badaf8cc2c","Type":"ContainerStarted","Data":"86bed8d80be6477f9c2d39525607b828fb191f84f25512bf9befecb1c32056d6"} Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.053063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ee4af418-0d2a-4797-9e6d-00badaf8cc2c","Type":"ContainerStarted","Data":"6275940742e3d78e959d8025467e72053df5fd46d225448291b3f598dbf9d72a"} Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.054925 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.055117 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.555084931 +0000 UTC m=+143.827542283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.055196 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-catalog-content\") pod \"community-operators-hrw24\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.055252 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-utilities\") pod \"community-operators-hrw24\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.055305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.055337 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8g96\" (UniqueName: \"kubernetes.io/projected/b1d9678d-171a-427b-9b37-192f523e63b2-kube-api-access-x8g96\") pod \"community-operators-hrw24\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.055712 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.555694348 +0000 UTC m=+143.828151690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.055758 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-catalog-content\") pod \"community-operators-hrw24\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.056232 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-utilities\") pod \"community-operators-hrw24\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.078532 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7hk4l"] Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.079512 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.087191 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.120492 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8g96\" (UniqueName: \"kubernetes.io/projected/b1d9678d-171a-427b-9b37-192f523e63b2-kube-api-access-x8g96\") pod \"community-operators-hrw24\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.124639 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.124607447 podStartE2EDuration="2.124607447s" podCreationTimestamp="2025-11-27 17:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:48.106223666 +0000 UTC m=+143.378681018" watchObservedRunningTime="2025-11-27 17:11:48.124607447 +0000 UTC m=+143.397064799" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.128480 4809 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.135875 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7hk4l"] Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.157358 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.157837 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-catalog-content\") pod \"certified-operators-7hk4l\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.157965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2qzw\" (UniqueName: \"kubernetes.io/projected/03562c92-cba5-4c9a-b759-7da40962be62-kube-api-access-v2qzw\") pod \"certified-operators-7hk4l\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.158028 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-utilities\") pod \"certified-operators-7hk4l\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.159028 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.658999069 +0000 UTC m=+143.931456421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.235851 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-98fbk"] Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.237164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.256877 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-98fbk"] Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.258882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-utilities\") pod \"certified-operators-7hk4l\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.258926 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgc92\" (UniqueName: \"kubernetes.io/projected/72f7df98-e370-4a9b-bde4-f5ebccd29aca-kube-api-access-wgc92\") pod \"community-operators-98fbk\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.258979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-catalog-content\") pod \"certified-operators-7hk4l\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.258999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-utilities\") pod \"community-operators-98fbk\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.259022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.259075 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2qzw\" (UniqueName: \"kubernetes.io/projected/03562c92-cba5-4c9a-b759-7da40962be62-kube-api-access-v2qzw\") pod \"certified-operators-7hk4l\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.259091 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-catalog-content\") pod \"community-operators-98fbk\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.259493 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-utilities\") pod \"certified-operators-7hk4l\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.259758 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-catalog-content\") pod \"certified-operators-7hk4l\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.260008 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.759995422 +0000 UTC m=+144.032452774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.287049 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2qzw\" (UniqueName: \"kubernetes.io/projected/03562c92-cba5-4c9a-b759-7da40962be62-kube-api-access-v2qzw\") pod \"certified-operators-7hk4l\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.360066 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.360290 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.860259514 +0000 UTC m=+144.132716866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.360392 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-utilities\") pod \"community-operators-98fbk\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.360431 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.360511 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-catalog-content\") pod \"community-operators-98fbk\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.360628 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgc92\" (UniqueName: \"kubernetes.io/projected/72f7df98-e370-4a9b-bde4-f5ebccd29aca-kube-api-access-wgc92\") pod \"community-operators-98fbk\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.360894 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-utilities\") pod \"community-operators-98fbk\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.360971 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-catalog-content\") pod \"community-operators-98fbk\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.361041 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.861018956 +0000 UTC m=+144.133476298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.382407 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgc92\" (UniqueName: \"kubernetes.io/projected/72f7df98-e370-4a9b-bde4-f5ebccd29aca-kube-api-access-wgc92\") pod \"community-operators-98fbk\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.394022 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.444813 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g6xz6"] Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.445663 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.461712 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.461911 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.961881335 +0000 UTC m=+144.234338697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.462072 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.462394 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:48.96238301 +0000 UTC m=+144.234840362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.477811 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g6xz6"] Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.563488 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.563995 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnp4t\" (UniqueName: \"kubernetes.io/projected/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-kube-api-access-fnp4t\") pod \"certified-operators-g6xz6\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.564085 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 17:11:49.064058173 +0000 UTC m=+144.336515535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.564300 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-catalog-content\") pod \"certified-operators-g6xz6\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.564406 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-utilities\") pod \"certified-operators-g6xz6\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.564448 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:48 crc kubenswrapper[4809]: E1127 17:11:48.565006 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 17:11:49.06499203 +0000 UTC m=+144.337449382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dj24l" (UID: "480a6028-0611-40c1-8204-0e2e37608800") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.623423 4809 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-27T17:11:48.128511549Z","Handler":null,"Name":""} Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.627442 4809 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.627491 4809 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.665680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.666572 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnp4t\" (UniqueName: \"kubernetes.io/projected/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-kube-api-access-fnp4t\") pod \"certified-operators-g6xz6\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.666634 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-catalog-content\") pod \"certified-operators-g6xz6\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.666663 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-utilities\") pod \"certified-operators-g6xz6\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.667411 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-utilities\") pod \"certified-operators-g6xz6\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.667718 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-catalog-content\") pod \"certified-operators-g6xz6\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.669631 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.691179 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnp4t\" (UniqueName: \"kubernetes.io/projected/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-kube-api-access-fnp4t\") pod \"certified-operators-g6xz6\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.726643 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7hk4l"] Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.765863 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.767862 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.771628 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.771836 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.807537 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dj24l\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.826210 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.856133 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 17:11:48 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 27 17:11:48 crc kubenswrapper[4809]: [+]process-running ok Nov 27 17:11:48 crc kubenswrapper[4809]: healthz check failed Nov 27 17:11:48 crc kubenswrapper[4809]: I1127 17:11:48.856217 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.059643 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g6xz6"] Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.061774 4809 generic.go:334] "Generic (PLEG): container finished" podID="15c3167d-77d3-457f-b602-5e87f96afe8d" containerID="3c3714c1a6319450a3525d33ff2a1a82e6ca523ba356cf684a9d85b33b861836" exitCode=0 Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.061867 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" event={"ID":"15c3167d-77d3-457f-b602-5e87f96afe8d","Type":"ContainerDied","Data":"3c3714c1a6319450a3525d33ff2a1a82e6ca523ba356cf684a9d85b33b861836"} Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.066834 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" event={"ID":"d6233aa1-8f36-44c3-a8c3-f20f1cfa4dc5","Type":"ContainerStarted","Data":"16948a67e01e614fc5a414622085b19b566db113ffce12b59d9b166ae4674490"} Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.073699 4809 generic.go:334] "Generic (PLEG): container finished" podID="ee4af418-0d2a-4797-9e6d-00badaf8cc2c" containerID="86bed8d80be6477f9c2d39525607b828fb191f84f25512bf9befecb1c32056d6" exitCode=0 Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.073926 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ee4af418-0d2a-4797-9e6d-00badaf8cc2c","Type":"ContainerDied","Data":"86bed8d80be6477f9c2d39525607b828fb191f84f25512bf9befecb1c32056d6"} Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.083790 4809 generic.go:334] "Generic (PLEG): container finished" podID="03562c92-cba5-4c9a-b759-7da40962be62" containerID="592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8" exitCode=0 Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.084107 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hk4l" event={"ID":"03562c92-cba5-4c9a-b759-7da40962be62","Type":"ContainerDied","Data":"592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8"} Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.084174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hk4l" event={"ID":"03562c92-cba5-4c9a-b759-7da40962be62","Type":"ContainerStarted","Data":"16d65a77c932312d8c53ee42bccedbc803f51f53d2bc3dec491a5924d98a8f20"} Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.089343 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.110293 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-dpfxv" podStartSLOduration=12.110268199 podStartE2EDuration="12.110268199s" podCreationTimestamp="2025-11-27 17:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:49.107206151 +0000 UTC m=+144.379663523" watchObservedRunningTime="2025-11-27 17:11:49.110268199 +0000 UTC m=+144.382725551" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.126624 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dj24l"] Nov 27 17:11:49 crc kubenswrapper[4809]: W1127 17:11:49.138021 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod480a6028_0611_40c1_8204_0e2e37608800.slice/crio-6cb3bf7ce5beb021d7744a931799e9f9397f8992cadaf92bd9ce73ab8acbb96f WatchSource:0}: Error finding container 6cb3bf7ce5beb021d7744a931799e9f9397f8992cadaf92bd9ce73ab8acbb96f: Status 404 returned error can't find the container with id 6cb3bf7ce5beb021d7744a931799e9f9397f8992cadaf92bd9ce73ab8acbb96f Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.183705 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/community-operators-hrw24" secret="" err="failed to sync secret cache: timed out waiting for the condition" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.183863 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.323936 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mqxzb" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.397223 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.409400 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.411000 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrw24"] Nov 27 17:11:49 crc kubenswrapper[4809]: W1127 17:11:49.423025 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1d9678d_171a_427b_9b37_192f523e63b2.slice/crio-bf9d146f95ef73133aca3a81ccf5559f8f6a24391f18a43aa1d9c1388a1497b9 WatchSource:0}: Error finding container bf9d146f95ef73133aca3a81ccf5559f8f6a24391f18a43aa1d9c1388a1497b9: Status 404 returned error can't find the container with id bf9d146f95ef73133aca3a81ccf5559f8f6a24391f18a43aa1d9c1388a1497b9 Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.467237 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.667288 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-98fbk"] Nov 27 17:11:49 crc kubenswrapper[4809]: W1127 17:11:49.683069 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72f7df98_e370_4a9b_bde4_f5ebccd29aca.slice/crio-71f95ff2b52e54a226171840d8e5a2bf0513296bb7a31c591cb3dde1ea36a165 WatchSource:0}: Error finding container 71f95ff2b52e54a226171840d8e5a2bf0513296bb7a31c591cb3dde1ea36a165: Status 404 returned error can't find the container with id 71f95ff2b52e54a226171840d8e5a2bf0513296bb7a31c591cb3dde1ea36a165 Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.823434 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.831318 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wpq4q" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.861452 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 17:11:49 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 27 17:11:49 crc kubenswrapper[4809]: [+]process-running ok Nov 27 17:11:49 crc kubenswrapper[4809]: healthz check failed Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.861578 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.899545 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.899941 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.910233 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.952863 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.952965 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.958533 4809 patch_prober.go:28] interesting pod/console-f9d7485db-w7f6q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.958604 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-w7f6q" podUID="5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.965885 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.966592 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.969593 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.969804 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.983000 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.992311 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8de1b04-296f-4ea2-bcad-af0f144dc274-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d8de1b04-296f-4ea2-bcad-af0f144dc274\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:49 crc kubenswrapper[4809]: I1127 17:11:49.992436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8de1b04-296f-4ea2-bcad-af0f144dc274-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d8de1b04-296f-4ea2-bcad-af0f144dc274\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.042176 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-jtxfb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.042298 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jtxfb" podUID="fff4c3fa-a124-4cb5-806d-eea27e5dfd4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.042322 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-jtxfb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.042436 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-jtxfb" podUID="fff4c3fa-a124-4cb5-806d-eea27e5dfd4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.042903 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-7c7gv" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.052425 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-chw9h"] Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.062890 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-chw9h"] Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.063087 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.068015 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.094096 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtnx6\" (UniqueName: \"kubernetes.io/projected/867f6bd0-b172-4409-a736-cfa7484d5685-kube-api-access-gtnx6\") pod \"redhat-marketplace-chw9h\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.094219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8de1b04-296f-4ea2-bcad-af0f144dc274-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d8de1b04-296f-4ea2-bcad-af0f144dc274\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.094284 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-utilities\") pod \"redhat-marketplace-chw9h\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.094316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8de1b04-296f-4ea2-bcad-af0f144dc274-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d8de1b04-296f-4ea2-bcad-af0f144dc274\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.094384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-catalog-content\") pod \"redhat-marketplace-chw9h\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.095039 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8de1b04-296f-4ea2-bcad-af0f144dc274-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d8de1b04-296f-4ea2-bcad-af0f144dc274\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.111555 4809 generic.go:334] "Generic (PLEG): container finished" podID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerID="0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138" exitCode=0 Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.111672 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbk" event={"ID":"72f7df98-e370-4a9b-bde4-f5ebccd29aca","Type":"ContainerDied","Data":"0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138"} Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.111706 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbk" event={"ID":"72f7df98-e370-4a9b-bde4-f5ebccd29aca","Type":"ContainerStarted","Data":"71f95ff2b52e54a226171840d8e5a2bf0513296bb7a31c591cb3dde1ea36a165"} Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.114938 4809 generic.go:334] "Generic (PLEG): container finished" podID="b1d9678d-171a-427b-9b37-192f523e63b2" containerID="14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca" exitCode=0 Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.115240 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrw24" event={"ID":"b1d9678d-171a-427b-9b37-192f523e63b2","Type":"ContainerDied","Data":"14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca"} Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.115347 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrw24" event={"ID":"b1d9678d-171a-427b-9b37-192f523e63b2","Type":"ContainerStarted","Data":"bf9d146f95ef73133aca3a81ccf5559f8f6a24391f18a43aa1d9c1388a1497b9"} Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.127157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" event={"ID":"480a6028-0611-40c1-8204-0e2e37608800","Type":"ContainerStarted","Data":"49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349"} Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.127767 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" event={"ID":"480a6028-0611-40c1-8204-0e2e37608800","Type":"ContainerStarted","Data":"6cb3bf7ce5beb021d7744a931799e9f9397f8992cadaf92bd9ce73ab8acbb96f"} Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.129003 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.133876 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g6xz6" event={"ID":"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83","Type":"ContainerDied","Data":"adc31ca2caefd2101822c3d9be2bc29282b164b3ea9ecb913ed6f44f97ad4cd9"} Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.136392 4809 generic.go:334] "Generic (PLEG): container finished" podID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerID="adc31ca2caefd2101822c3d9be2bc29282b164b3ea9ecb913ed6f44f97ad4cd9" exitCode=0 Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.141393 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8de1b04-296f-4ea2-bcad-af0f144dc274-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d8de1b04-296f-4ea2-bcad-af0f144dc274\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.141495 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g6xz6" event={"ID":"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83","Type":"ContainerStarted","Data":"73182e9db74c63d149e63f73f4252a87fbbeb1c3f45602c173fb8889cd1e6463"} Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.164868 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-hgmrs" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.167726 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" podStartSLOduration=125.167702681 podStartE2EDuration="2m5.167702681s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:11:50.158188727 +0000 UTC m=+145.430646079" watchObservedRunningTime="2025-11-27 17:11:50.167702681 +0000 UTC m=+145.440160033" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.196477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-utilities\") pod \"redhat-marketplace-chw9h\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.196608 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-catalog-content\") pod \"redhat-marketplace-chw9h\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.196758 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtnx6\" (UniqueName: \"kubernetes.io/projected/867f6bd0-b172-4409-a736-cfa7484d5685-kube-api-access-gtnx6\") pod \"redhat-marketplace-chw9h\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.198497 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-utilities\") pod \"redhat-marketplace-chw9h\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.202210 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-catalog-content\") pod \"redhat-marketplace-chw9h\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.224223 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtnx6\" (UniqueName: \"kubernetes.io/projected/867f6bd0-b172-4409-a736-cfa7484d5685-kube-api-access-gtnx6\") pod \"redhat-marketplace-chw9h\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.298617 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.398682 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.435248 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xrfjz"] Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.436395 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.450426 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrfjz"] Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.501444 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-catalog-content\") pod \"redhat-marketplace-xrfjz\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.501942 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llqvm\" (UniqueName: \"kubernetes.io/projected/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-kube-api-access-llqvm\") pod \"redhat-marketplace-xrfjz\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.501979 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-utilities\") pod \"redhat-marketplace-xrfjz\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.536703 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.559239 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.602424 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kubelet-dir\") pod \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\" (UID: \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\") " Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.602533 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv98z\" (UniqueName: \"kubernetes.io/projected/15c3167d-77d3-457f-b602-5e87f96afe8d-kube-api-access-tv98z\") pod \"15c3167d-77d3-457f-b602-5e87f96afe8d\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.602586 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15c3167d-77d3-457f-b602-5e87f96afe8d-config-volume\") pod \"15c3167d-77d3-457f-b602-5e87f96afe8d\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.602609 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15c3167d-77d3-457f-b602-5e87f96afe8d-secret-volume\") pod \"15c3167d-77d3-457f-b602-5e87f96afe8d\" (UID: \"15c3167d-77d3-457f-b602-5e87f96afe8d\") " Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.602626 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kube-api-access\") pod \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\" (UID: \"ee4af418-0d2a-4797-9e6d-00badaf8cc2c\") " Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.602825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-catalog-content\") pod \"redhat-marketplace-xrfjz\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.602848 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llqvm\" (UniqueName: \"kubernetes.io/projected/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-kube-api-access-llqvm\") pod \"redhat-marketplace-xrfjz\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.602882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-utilities\") pod \"redhat-marketplace-xrfjz\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.603963 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-utilities\") pod \"redhat-marketplace-xrfjz\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.604014 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ee4af418-0d2a-4797-9e6d-00badaf8cc2c" (UID: "ee4af418-0d2a-4797-9e6d-00badaf8cc2c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.605704 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c3167d-77d3-457f-b602-5e87f96afe8d-config-volume" (OuterVolumeSpecName: "config-volume") pod "15c3167d-77d3-457f-b602-5e87f96afe8d" (UID: "15c3167d-77d3-457f-b602-5e87f96afe8d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.606033 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-catalog-content\") pod \"redhat-marketplace-xrfjz\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.613456 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ee4af418-0d2a-4797-9e6d-00badaf8cc2c" (UID: "ee4af418-0d2a-4797-9e6d-00badaf8cc2c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.614699 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15c3167d-77d3-457f-b602-5e87f96afe8d-kube-api-access-tv98z" (OuterVolumeSpecName: "kube-api-access-tv98z") pod "15c3167d-77d3-457f-b602-5e87f96afe8d" (UID: "15c3167d-77d3-457f-b602-5e87f96afe8d"). InnerVolumeSpecName "kube-api-access-tv98z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.614847 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15c3167d-77d3-457f-b602-5e87f96afe8d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "15c3167d-77d3-457f-b602-5e87f96afe8d" (UID: "15c3167d-77d3-457f-b602-5e87f96afe8d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.627782 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llqvm\" (UniqueName: \"kubernetes.io/projected/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-kube-api-access-llqvm\") pod \"redhat-marketplace-xrfjz\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.708843 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv98z\" (UniqueName: \"kubernetes.io/projected/15c3167d-77d3-457f-b602-5e87f96afe8d-kube-api-access-tv98z\") on node \"crc\" DevicePath \"\"" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.709221 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15c3167d-77d3-457f-b602-5e87f96afe8d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.709232 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15c3167d-77d3-457f-b602-5e87f96afe8d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.709242 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.709253 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee4af418-0d2a-4797-9e6d-00badaf8cc2c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.720131 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-chw9h"] Nov 27 17:11:50 crc kubenswrapper[4809]: W1127 17:11:50.742477 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod867f6bd0_b172_4409_a736_cfa7484d5685.slice/crio-50b824330a9c147fc0aa98c1ded4ef065546b1504561abcf6c9b11beb76e0685 WatchSource:0}: Error finding container 50b824330a9c147fc0aa98c1ded4ef065546b1504561abcf6c9b11beb76e0685: Status 404 returned error can't find the container with id 50b824330a9c147fc0aa98c1ded4ef065546b1504561abcf6c9b11beb76e0685 Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.784633 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.807198 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.823185 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjg9h" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.836034 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qt7fm" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.855420 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.856717 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j6ggx" Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.866956 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.863984 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 17:11:50 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 27 17:11:50 crc kubenswrapper[4809]: [+]process-running ok Nov 27 17:11:50 crc kubenswrapper[4809]: healthz check failed Nov 27 17:11:50 crc kubenswrapper[4809]: I1127 17:11:50.867864 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 17:11:50 crc kubenswrapper[4809]: W1127 17:11:50.900317 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd8de1b04_296f_4ea2_bcad_af0f144dc274.slice/crio-63ef58538728482b12b3048a8b92e3d5adacb0cc119e4d3e9769794803b67a1c WatchSource:0}: Error finding container 63ef58538728482b12b3048a8b92e3d5adacb0cc119e4d3e9769794803b67a1c: Status 404 returned error can't find the container with id 63ef58538728482b12b3048a8b92e3d5adacb0cc119e4d3e9769794803b67a1c Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.047712 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dcvzh"] Nov 27 17:11:51 crc kubenswrapper[4809]: E1127 17:11:51.048461 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4af418-0d2a-4797-9e6d-00badaf8cc2c" containerName="pruner" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.048475 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4af418-0d2a-4797-9e6d-00badaf8cc2c" containerName="pruner" Nov 27 17:11:51 crc kubenswrapper[4809]: E1127 17:11:51.048491 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c3167d-77d3-457f-b602-5e87f96afe8d" containerName="collect-profiles" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.048497 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c3167d-77d3-457f-b602-5e87f96afe8d" containerName="collect-profiles" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.048588 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c3167d-77d3-457f-b602-5e87f96afe8d" containerName="collect-profiles" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.048606 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4af418-0d2a-4797-9e6d-00badaf8cc2c" containerName="pruner" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.053874 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.068115 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.078509 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dcvzh"] Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.140291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-catalog-content\") pod \"redhat-operators-dcvzh\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.140373 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-utilities\") pod \"redhat-operators-dcvzh\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.140402 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2776\" (UniqueName: \"kubernetes.io/projected/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-kube-api-access-l2776\") pod \"redhat-operators-dcvzh\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.153938 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.197444 4809 generic.go:334] "Generic (PLEG): container finished" podID="867f6bd0-b172-4409-a736-cfa7484d5685" containerID="a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a" exitCode=0 Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.197835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chw9h" event={"ID":"867f6bd0-b172-4409-a736-cfa7484d5685","Type":"ContainerDied","Data":"a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a"} Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.197921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chw9h" event={"ID":"867f6bd0-b172-4409-a736-cfa7484d5685","Type":"ContainerStarted","Data":"50b824330a9c147fc0aa98c1ded4ef065546b1504561abcf6c9b11beb76e0685"} Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.215063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" event={"ID":"15c3167d-77d3-457f-b602-5e87f96afe8d","Type":"ContainerDied","Data":"259e9d4bbad6c573d5ccac10badb7f2c191140864b5825fcef72c719c50f449d"} Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.215107 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="259e9d4bbad6c573d5ccac10badb7f2c191140864b5825fcef72c719c50f449d" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.215208 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.239066 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d8de1b04-296f-4ea2-bcad-af0f144dc274","Type":"ContainerStarted","Data":"63ef58538728482b12b3048a8b92e3d5adacb0cc119e4d3e9769794803b67a1c"} Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.241575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-catalog-content\") pod \"redhat-operators-dcvzh\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.241661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-utilities\") pod \"redhat-operators-dcvzh\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.241683 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2776\" (UniqueName: \"kubernetes.io/projected/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-kube-api-access-l2776\") pod \"redhat-operators-dcvzh\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.242903 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-utilities\") pod \"redhat-operators-dcvzh\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.243120 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-catalog-content\") pod \"redhat-operators-dcvzh\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.251861 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ee4af418-0d2a-4797-9e6d-00badaf8cc2c","Type":"ContainerDied","Data":"6275940742e3d78e959d8025467e72053df5fd46d225448291b3f598dbf9d72a"} Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.251941 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6275940742e3d78e959d8025467e72053df5fd46d225448291b3f598dbf9d72a" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.252081 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.296068 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2776\" (UniqueName: \"kubernetes.io/projected/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-kube-api-access-l2776\") pod \"redhat-operators-dcvzh\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.390269 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.479015 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jsvqm"] Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.480049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.484588 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jsvqm"] Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.598371 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrfjz"] Nov 27 17:11:51 crc kubenswrapper[4809]: W1127 17:11:51.631915 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b2f80bd_419e_4d4a_b4b5_e6e7362a20a7.slice/crio-d6c4daee56ddfb0ec1d5f43c6be9665c1ec06559de780c21cd1019a36cacddaf WatchSource:0}: Error finding container d6c4daee56ddfb0ec1d5f43c6be9665c1ec06559de780c21cd1019a36cacddaf: Status 404 returned error can't find the container with id d6c4daee56ddfb0ec1d5f43c6be9665c1ec06559de780c21cd1019a36cacddaf Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.658580 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-utilities\") pod \"redhat-operators-jsvqm\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.658650 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-catalog-content\") pod \"redhat-operators-jsvqm\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.659016 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nczqc\" (UniqueName: \"kubernetes.io/projected/824326f1-6377-4f05-a5fd-db94363ebe1d-kube-api-access-nczqc\") pod \"redhat-operators-jsvqm\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.721731 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dcvzh"] Nov 27 17:11:51 crc kubenswrapper[4809]: W1127 17:11:51.758238 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod225a59f5_d662_4f63_8ec6_c7f3eeb6ba40.slice/crio-f383b83a837a0b9f10fef1809bd65baccb1f83e237d9bc21db8c0b06cd6df181 WatchSource:0}: Error finding container f383b83a837a0b9f10fef1809bd65baccb1f83e237d9bc21db8c0b06cd6df181: Status 404 returned error can't find the container with id f383b83a837a0b9f10fef1809bd65baccb1f83e237d9bc21db8c0b06cd6df181 Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.760276 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-utilities\") pod \"redhat-operators-jsvqm\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.760319 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-catalog-content\") pod \"redhat-operators-jsvqm\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.760393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nczqc\" (UniqueName: \"kubernetes.io/projected/824326f1-6377-4f05-a5fd-db94363ebe1d-kube-api-access-nczqc\") pod \"redhat-operators-jsvqm\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.761152 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-catalog-content\") pod \"redhat-operators-jsvqm\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.761347 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-utilities\") pod \"redhat-operators-jsvqm\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.786805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nczqc\" (UniqueName: \"kubernetes.io/projected/824326f1-6377-4f05-a5fd-db94363ebe1d-kube-api-access-nczqc\") pod \"redhat-operators-jsvqm\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.813297 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.856453 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 17:11:51 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 27 17:11:51 crc kubenswrapper[4809]: [+]process-running ok Nov 27 17:11:51 crc kubenswrapper[4809]: healthz check failed Nov 27 17:11:51 crc kubenswrapper[4809]: I1127 17:11:51.856517 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.233566 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jsvqm"] Nov 27 17:11:52 crc kubenswrapper[4809]: W1127 17:11:52.265607 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod824326f1_6377_4f05_a5fd_db94363ebe1d.slice/crio-30b810804d431d1b4daf51edfd5ebaa4814f064b198a307c7f34fb0ebce2d4d7 WatchSource:0}: Error finding container 30b810804d431d1b4daf51edfd5ebaa4814f064b198a307c7f34fb0ebce2d4d7: Status 404 returned error can't find the container with id 30b810804d431d1b4daf51edfd5ebaa4814f064b198a307c7f34fb0ebce2d4d7 Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.270331 4809 generic.go:334] "Generic (PLEG): container finished" podID="d8de1b04-296f-4ea2-bcad-af0f144dc274" containerID="e4689a26943cc07fd3776b992e3918a3ce8b0d461ba3238276c680e69ae47821" exitCode=0 Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.270419 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d8de1b04-296f-4ea2-bcad-af0f144dc274","Type":"ContainerDied","Data":"e4689a26943cc07fd3776b992e3918a3ce8b0d461ba3238276c680e69ae47821"} Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.283818 4809 generic.go:334] "Generic (PLEG): container finished" podID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerID="34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6" exitCode=0 Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.284251 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcvzh" event={"ID":"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40","Type":"ContainerDied","Data":"34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6"} Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.284309 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcvzh" event={"ID":"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40","Type":"ContainerStarted","Data":"f383b83a837a0b9f10fef1809bd65baccb1f83e237d9bc21db8c0b06cd6df181"} Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.293702 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerID="4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4" exitCode=0 Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.293771 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrfjz" event={"ID":"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7","Type":"ContainerDied","Data":"4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4"} Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.293810 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrfjz" event={"ID":"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7","Type":"ContainerStarted","Data":"d6c4daee56ddfb0ec1d5f43c6be9665c1ec06559de780c21cd1019a36cacddaf"} Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.855318 4809 patch_prober.go:28] interesting pod/router-default-5444994796-kzdxz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 17:11:52 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Nov 27 17:11:52 crc kubenswrapper[4809]: [+]process-running ok Nov 27 17:11:52 crc kubenswrapper[4809]: healthz check failed Nov 27 17:11:52 crc kubenswrapper[4809]: I1127 17:11:52.856073 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kzdxz" podUID="9a3cce9e-247e-457d-b41c-a1a7ee3c7888" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.311328 4809 generic.go:334] "Generic (PLEG): container finished" podID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerID="933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554" exitCode=0 Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.311454 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsvqm" event={"ID":"824326f1-6377-4f05-a5fd-db94363ebe1d","Type":"ContainerDied","Data":"933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554"} Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.311535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsvqm" event={"ID":"824326f1-6377-4f05-a5fd-db94363ebe1d","Type":"ContainerStarted","Data":"30b810804d431d1b4daf51edfd5ebaa4814f064b198a307c7f34fb0ebce2d4d7"} Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.599358 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.599417 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.599443 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.603876 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.625222 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.627696 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.700631 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.705478 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.708445 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.720496 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.733449 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.802096 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8de1b04-296f-4ea2-bcad-af0f144dc274-kube-api-access\") pod \"d8de1b04-296f-4ea2-bcad-af0f144dc274\" (UID: \"d8de1b04-296f-4ea2-bcad-af0f144dc274\") " Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.802130 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8de1b04-296f-4ea2-bcad-af0f144dc274-kubelet-dir\") pod \"d8de1b04-296f-4ea2-bcad-af0f144dc274\" (UID: \"d8de1b04-296f-4ea2-bcad-af0f144dc274\") " Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.802371 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8de1b04-296f-4ea2-bcad-af0f144dc274-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d8de1b04-296f-4ea2-bcad-af0f144dc274" (UID: "d8de1b04-296f-4ea2-bcad-af0f144dc274"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.812632 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8de1b04-296f-4ea2-bcad-af0f144dc274-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d8de1b04-296f-4ea2-bcad-af0f144dc274" (UID: "d8de1b04-296f-4ea2-bcad-af0f144dc274"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.853631 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.860922 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-kzdxz" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.904305 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8de1b04-296f-4ea2-bcad-af0f144dc274-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.904339 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8de1b04-296f-4ea2-bcad-af0f144dc274-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:11:53 crc kubenswrapper[4809]: I1127 17:11:53.987088 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 17:11:54 crc kubenswrapper[4809]: W1127 17:11:54.272026 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-92ef7bd834e24dbe7f8af03512dfe781a0c3937e591cbfc556c5f1b40eebcea4 WatchSource:0}: Error finding container 92ef7bd834e24dbe7f8af03512dfe781a0c3937e591cbfc556c5f1b40eebcea4: Status 404 returned error can't find the container with id 92ef7bd834e24dbe7f8af03512dfe781a0c3937e591cbfc556c5f1b40eebcea4 Nov 27 17:11:54 crc kubenswrapper[4809]: I1127 17:11:54.375102 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"92ef7bd834e24dbe7f8af03512dfe781a0c3937e591cbfc556c5f1b40eebcea4"} Nov 27 17:11:54 crc kubenswrapper[4809]: I1127 17:11:54.384465 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 17:11:54 crc kubenswrapper[4809]: I1127 17:11:54.385788 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d8de1b04-296f-4ea2-bcad-af0f144dc274","Type":"ContainerDied","Data":"63ef58538728482b12b3048a8b92e3d5adacb0cc119e4d3e9769794803b67a1c"} Nov 27 17:11:54 crc kubenswrapper[4809]: I1127 17:11:54.385832 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63ef58538728482b12b3048a8b92e3d5adacb0cc119e4d3e9769794803b67a1c" Nov 27 17:11:54 crc kubenswrapper[4809]: W1127 17:11:54.441387 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-7cdf9f0ae39126983834092c3e054f300d981cd81c418d6422115a25ea8e1454 WatchSource:0}: Error finding container 7cdf9f0ae39126983834092c3e054f300d981cd81c418d6422115a25ea8e1454: Status 404 returned error can't find the container with id 7cdf9f0ae39126983834092c3e054f300d981cd81c418d6422115a25ea8e1454 Nov 27 17:11:54 crc kubenswrapper[4809]: W1127 17:11:54.665021 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-0edefb76344358abf9858d15e72402a105e9ed713b5471a030f532bf20630793 WatchSource:0}: Error finding container 0edefb76344358abf9858d15e72402a105e9ed713b5471a030f532bf20630793: Status 404 returned error can't find the container with id 0edefb76344358abf9858d15e72402a105e9ed713b5471a030f532bf20630793 Nov 27 17:11:54 crc kubenswrapper[4809]: I1127 17:11:54.795028 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:11:55 crc kubenswrapper[4809]: I1127 17:11:55.426508 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9e48de387a58dfdc1a18a700d03d5055f5936ecd234bca7ba40868cb5b9e80f2"} Nov 27 17:11:55 crc kubenswrapper[4809]: I1127 17:11:55.426924 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7cdf9f0ae39126983834092c3e054f300d981cd81c418d6422115a25ea8e1454"} Nov 27 17:11:55 crc kubenswrapper[4809]: I1127 17:11:55.427308 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:11:55 crc kubenswrapper[4809]: I1127 17:11:55.448129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d41a5d2bfbd200903ec8649ff857c6a6a18340f114dfc80519c7db9c7617b18e"} Nov 27 17:11:55 crc kubenswrapper[4809]: I1127 17:11:55.448942 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0edefb76344358abf9858d15e72402a105e9ed713b5471a030f532bf20630793"} Nov 27 17:11:55 crc kubenswrapper[4809]: I1127 17:11:55.506494 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"99bb2cc74e9230c527806f629efc71a26b46f05d432627b936652fead873ee83"} Nov 27 17:11:55 crc kubenswrapper[4809]: I1127 17:11:55.781666 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:11:55 crc kubenswrapper[4809]: I1127 17:11:55.781718 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:11:55 crc kubenswrapper[4809]: I1127 17:11:55.910599 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6v9pw" Nov 27 17:11:59 crc kubenswrapper[4809]: I1127 17:11:59.958945 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:11:59 crc kubenswrapper[4809]: I1127 17:11:59.964964 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:12:00 crc kubenswrapper[4809]: I1127 17:12:00.056656 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-jtxfb" Nov 27 17:12:06 crc kubenswrapper[4809]: I1127 17:12:06.936597 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:12:06 crc kubenswrapper[4809]: I1127 17:12:06.944937 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/683f4d0e-67aa-40a7-b7b0-e5325ed0225f-metrics-certs\") pod \"network-metrics-daemon-2nc2d\" (UID: \"683f4d0e-67aa-40a7-b7b0-e5325ed0225f\") " pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:12:07 crc kubenswrapper[4809]: I1127 17:12:07.204329 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nc2d" Nov 27 17:12:07 crc kubenswrapper[4809]: I1127 17:12:07.441332 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2nc2d"] Nov 27 17:12:07 crc kubenswrapper[4809]: I1127 17:12:07.711957 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" event={"ID":"683f4d0e-67aa-40a7-b7b0-e5325ed0225f","Type":"ContainerStarted","Data":"bbd14a242fb7a8164b4d82afa47c291483705192845de26aecfa315b21b2f6b3"} Nov 27 17:12:08 crc kubenswrapper[4809]: I1127 17:12:08.836150 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:12:09 crc kubenswrapper[4809]: I1127 17:12:09.727280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" event={"ID":"683f4d0e-67aa-40a7-b7b0-e5325ed0225f","Type":"ContainerStarted","Data":"d8d17b8bee5cc93a699b125b0cac910082487b1c2dde953181a85e0a82e00ba6"} Nov 27 17:12:20 crc kubenswrapper[4809]: I1127 17:12:20.819709 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rsqzp" Nov 27 17:12:20 crc kubenswrapper[4809]: E1127 17:12:20.952374 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 27 17:12:20 crc kubenswrapper[4809]: E1127 17:12:20.952724 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gtnx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-chw9h_openshift-marketplace(867f6bd0-b172-4409-a736-cfa7484d5685): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 17:12:20 crc kubenswrapper[4809]: E1127 17:12:20.954900 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-chw9h" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" Nov 27 17:12:23 crc kubenswrapper[4809]: I1127 17:12:23.984670 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 17:12:23 crc kubenswrapper[4809]: E1127 17:12:23.986452 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8de1b04-296f-4ea2-bcad-af0f144dc274" containerName="pruner" Nov 27 17:12:23 crc kubenswrapper[4809]: I1127 17:12:23.986470 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8de1b04-296f-4ea2-bcad-af0f144dc274" containerName="pruner" Nov 27 17:12:23 crc kubenswrapper[4809]: I1127 17:12:23.988369 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8de1b04-296f-4ea2-bcad-af0f144dc274" containerName="pruner" Nov 27 17:12:23 crc kubenswrapper[4809]: I1127 17:12:23.989265 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:23 crc kubenswrapper[4809]: I1127 17:12:23.992923 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 17:12:23 crc kubenswrapper[4809]: I1127 17:12:23.993100 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 17:12:23 crc kubenswrapper[4809]: I1127 17:12:23.998428 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 17:12:24 crc kubenswrapper[4809]: I1127 17:12:24.109668 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:24 crc kubenswrapper[4809]: I1127 17:12:24.110144 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:24 crc kubenswrapper[4809]: I1127 17:12:24.211716 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:24 crc kubenswrapper[4809]: I1127 17:12:24.211841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:24 crc kubenswrapper[4809]: I1127 17:12:24.211928 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:24 crc kubenswrapper[4809]: I1127 17:12:24.234646 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:24 crc kubenswrapper[4809]: I1127 17:12:24.311509 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:24 crc kubenswrapper[4809]: E1127 17:12:24.677660 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-chw9h" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" Nov 27 17:12:24 crc kubenswrapper[4809]: E1127 17:12:24.765600 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 27 17:12:24 crc kubenswrapper[4809]: E1127 17:12:24.765795 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l2776,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-dcvzh_openshift-marketplace(225a59f5-d662-4f63-8ec6-c7f3eeb6ba40): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 17:12:24 crc kubenswrapper[4809]: E1127 17:12:24.767146 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-dcvzh" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" Nov 27 17:12:25 crc kubenswrapper[4809]: I1127 17:12:25.779521 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:12:25 crc kubenswrapper[4809]: I1127 17:12:25.779966 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:12:26 crc kubenswrapper[4809]: E1127 17:12:26.271095 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-dcvzh" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" Nov 27 17:12:26 crc kubenswrapper[4809]: E1127 17:12:26.332048 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 27 17:12:26 crc kubenswrapper[4809]: E1127 17:12:26.332242 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wgc92,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-98fbk_openshift-marketplace(72f7df98-e370-4a9b-bde4-f5ebccd29aca): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 17:12:26 crc kubenswrapper[4809]: E1127 17:12:26.333626 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-98fbk" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" Nov 27 17:12:26 crc kubenswrapper[4809]: E1127 17:12:26.404472 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 27 17:12:26 crc kubenswrapper[4809]: E1127 17:12:26.404648 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x8g96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hrw24_openshift-marketplace(b1d9678d-171a-427b-9b37-192f523e63b2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 17:12:26 crc kubenswrapper[4809]: E1127 17:12:26.406211 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hrw24" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.668938 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hrw24" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.668974 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-98fbk" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.747301 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.747487 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v2qzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7hk4l_openshift-marketplace(03562c92-cba5-4c9a-b759-7da40962be62): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.749490 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7hk4l" podUID="03562c92-cba5-4c9a-b759-7da40962be62" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.760015 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.760247 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-llqvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xrfjz_openshift-marketplace(6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.761510 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xrfjz" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.809160 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.809338 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fnp4t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-g6xz6_openshift-marketplace(289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.810641 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-g6xz6" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.884864 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7hk4l" podUID="03562c92-cba5-4c9a-b759-7da40962be62" Nov 27 17:12:27 crc kubenswrapper[4809]: E1127 17:12:27.891012 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-g6xz6" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" Nov 27 17:12:27 crc kubenswrapper[4809]: I1127 17:12:27.958579 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 17:12:28 crc kubenswrapper[4809]: I1127 17:12:28.893797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2nc2d" event={"ID":"683f4d0e-67aa-40a7-b7b0-e5325ed0225f","Type":"ContainerStarted","Data":"d7a3ed0e673be65d5b59476a6e33c0e51918a910f1a1bc1a6804f618d1720beb"} Nov 27 17:12:28 crc kubenswrapper[4809]: I1127 17:12:28.903330 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"db95db91-9ee8-4b4b-b823-bb0b7a03a56c","Type":"ContainerStarted","Data":"9fa7f4dab3453c51ed62092f278229c3bfd43f03d1a4d8683005daeaaa0488a5"} Nov 27 17:12:28 crc kubenswrapper[4809]: I1127 17:12:28.903418 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"db95db91-9ee8-4b4b-b823-bb0b7a03a56c","Type":"ContainerStarted","Data":"f51f5ab4cb53653d7feb65e71f42cec391b56a040ee9f3db300666d7de7b23b7"} Nov 27 17:12:28 crc kubenswrapper[4809]: I1127 17:12:28.919419 4809 generic.go:334] "Generic (PLEG): container finished" podID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerID="0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624" exitCode=0 Nov 27 17:12:28 crc kubenswrapper[4809]: I1127 17:12:28.919500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsvqm" event={"ID":"824326f1-6377-4f05-a5fd-db94363ebe1d","Type":"ContainerDied","Data":"0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624"} Nov 27 17:12:28 crc kubenswrapper[4809]: I1127 17:12:28.955381 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-2nc2d" podStartSLOduration=163.955355845 podStartE2EDuration="2m43.955355845s" podCreationTimestamp="2025-11-27 17:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:12:28.932864737 +0000 UTC m=+184.205322089" watchObservedRunningTime="2025-11-27 17:12:28.955355845 +0000 UTC m=+184.227813197" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.180471 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=6.180449008 podStartE2EDuration="6.180449008s" podCreationTimestamp="2025-11-27 17:12:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:12:28.972772088 +0000 UTC m=+184.245229440" watchObservedRunningTime="2025-11-27 17:12:29.180449008 +0000 UTC m=+184.452906360" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.181002 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.181648 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.195050 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.285153 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-var-lock\") pod \"installer-9-crc\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.285239 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kube-api-access\") pod \"installer-9-crc\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.285445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.387114 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kube-api-access\") pod \"installer-9-crc\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.387331 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.387449 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-var-lock\") pod \"installer-9-crc\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.387450 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.387560 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-var-lock\") pod \"installer-9-crc\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.409374 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kube-api-access\") pod \"installer-9-crc\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.512971 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.707832 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.929917 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsvqm" event={"ID":"824326f1-6377-4f05-a5fd-db94363ebe1d","Type":"ContainerStarted","Data":"31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f"} Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.932321 4809 generic.go:334] "Generic (PLEG): container finished" podID="db95db91-9ee8-4b4b-b823-bb0b7a03a56c" containerID="9fa7f4dab3453c51ed62092f278229c3bfd43f03d1a4d8683005daeaaa0488a5" exitCode=0 Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.932386 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"db95db91-9ee8-4b4b-b823-bb0b7a03a56c","Type":"ContainerDied","Data":"9fa7f4dab3453c51ed62092f278229c3bfd43f03d1a4d8683005daeaaa0488a5"} Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.935072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9","Type":"ContainerStarted","Data":"07005d6a3ef8d44153eb4b1b15242282faacad1d80e2c3702140177195b95559"} Nov 27 17:12:29 crc kubenswrapper[4809]: I1127 17:12:29.963190 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jsvqm" podStartSLOduration=2.852647145 podStartE2EDuration="38.963168697s" podCreationTimestamp="2025-11-27 17:11:51 +0000 UTC" firstStartedPulling="2025-11-27 17:11:53.314464871 +0000 UTC m=+148.586922223" lastFinishedPulling="2025-11-27 17:12:29.424986423 +0000 UTC m=+184.697443775" observedRunningTime="2025-11-27 17:12:29.957160753 +0000 UTC m=+185.229618115" watchObservedRunningTime="2025-11-27 17:12:29.963168697 +0000 UTC m=+185.235626069" Nov 27 17:12:30 crc kubenswrapper[4809]: I1127 17:12:30.944045 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9","Type":"ContainerStarted","Data":"467e6364808ccaab2258fb7ca72a66e4cc15eb93b844720a9319493d61baccaf"} Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.178913 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.201017 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.200984052 podStartE2EDuration="2.200984052s" podCreationTimestamp="2025-11-27 17:12:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:12:30.970161743 +0000 UTC m=+186.242619095" watchObservedRunningTime="2025-11-27 17:12:31.200984052 +0000 UTC m=+186.473441404" Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.315016 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kubelet-dir\") pod \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\" (UID: \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\") " Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.315173 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kube-api-access\") pod \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\" (UID: \"db95db91-9ee8-4b4b-b823-bb0b7a03a56c\") " Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.315530 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "db95db91-9ee8-4b4b-b823-bb0b7a03a56c" (UID: "db95db91-9ee8-4b4b-b823-bb0b7a03a56c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.323772 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "db95db91-9ee8-4b4b-b823-bb0b7a03a56c" (UID: "db95db91-9ee8-4b4b-b823-bb0b7a03a56c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.416447 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.416485 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db95db91-9ee8-4b4b-b823-bb0b7a03a56c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.814761 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.815166 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.948535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"db95db91-9ee8-4b4b-b823-bb0b7a03a56c","Type":"ContainerDied","Data":"f51f5ab4cb53653d7feb65e71f42cec391b56a040ee9f3db300666d7de7b23b7"} Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.948593 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f51f5ab4cb53653d7feb65e71f42cec391b56a040ee9f3db300666d7de7b23b7" Nov 27 17:12:31 crc kubenswrapper[4809]: I1127 17:12:31.948662 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 17:12:32 crc kubenswrapper[4809]: I1127 17:12:32.938805 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jsvqm" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerName="registry-server" probeResult="failure" output=< Nov 27 17:12:32 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 27 17:12:32 crc kubenswrapper[4809]: > Nov 27 17:12:33 crc kubenswrapper[4809]: I1127 17:12:33.727221 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 17:12:39 crc kubenswrapper[4809]: I1127 17:12:39.398895 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2t2fb"] Nov 27 17:12:41 crc kubenswrapper[4809]: I1127 17:12:41.875418 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:12:41 crc kubenswrapper[4809]: I1127 17:12:41.919018 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:12:42 crc kubenswrapper[4809]: I1127 17:12:42.008504 4809 generic.go:334] "Generic (PLEG): container finished" podID="867f6bd0-b172-4409-a736-cfa7484d5685" containerID="1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66" exitCode=0 Nov 27 17:12:42 crc kubenswrapper[4809]: I1127 17:12:42.008590 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chw9h" event={"ID":"867f6bd0-b172-4409-a736-cfa7484d5685","Type":"ContainerDied","Data":"1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66"} Nov 27 17:12:42 crc kubenswrapper[4809]: I1127 17:12:42.012192 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcvzh" event={"ID":"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40","Type":"ContainerStarted","Data":"871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793"} Nov 27 17:12:42 crc kubenswrapper[4809]: I1127 17:12:42.029844 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerID="1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243" exitCode=0 Nov 27 17:12:42 crc kubenswrapper[4809]: I1127 17:12:42.029914 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrfjz" event={"ID":"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7","Type":"ContainerDied","Data":"1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243"} Nov 27 17:12:42 crc kubenswrapper[4809]: I1127 17:12:42.032838 4809 generic.go:334] "Generic (PLEG): container finished" podID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerID="27dff38a6eeee8d54ee2a38c30f447edae3c769c14a5ca5714965a1e327c2469" exitCode=0 Nov 27 17:12:42 crc kubenswrapper[4809]: I1127 17:12:42.032926 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g6xz6" event={"ID":"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83","Type":"ContainerDied","Data":"27dff38a6eeee8d54ee2a38c30f447edae3c769c14a5ca5714965a1e327c2469"} Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.040414 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chw9h" event={"ID":"867f6bd0-b172-4409-a736-cfa7484d5685","Type":"ContainerStarted","Data":"1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28"} Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.041995 4809 generic.go:334] "Generic (PLEG): container finished" podID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerID="871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793" exitCode=0 Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.042053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcvzh" event={"ID":"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40","Type":"ContainerDied","Data":"871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793"} Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.044397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrfjz" event={"ID":"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7","Type":"ContainerStarted","Data":"0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017"} Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.047530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g6xz6" event={"ID":"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83","Type":"ContainerStarted","Data":"11bb3554f0667ce724c0695fb33793d3db6381f2f8da4acb48b39daec19b9c04"} Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.051696 4809 generic.go:334] "Generic (PLEG): container finished" podID="03562c92-cba5-4c9a-b759-7da40962be62" containerID="7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f" exitCode=0 Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.051812 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hk4l" event={"ID":"03562c92-cba5-4c9a-b759-7da40962be62","Type":"ContainerDied","Data":"7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f"} Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.068983 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-chw9h" podStartSLOduration=1.509937169 podStartE2EDuration="53.068955401s" podCreationTimestamp="2025-11-27 17:11:50 +0000 UTC" firstStartedPulling="2025-11-27 17:11:51.202659715 +0000 UTC m=+146.475117067" lastFinishedPulling="2025-11-27 17:12:42.761677947 +0000 UTC m=+198.034135299" observedRunningTime="2025-11-27 17:12:43.065281545 +0000 UTC m=+198.337738897" watchObservedRunningTime="2025-11-27 17:12:43.068955401 +0000 UTC m=+198.341412753" Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.130370 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xrfjz" podStartSLOduration=3.739122611 podStartE2EDuration="53.130349352s" podCreationTimestamp="2025-11-27 17:11:50 +0000 UTC" firstStartedPulling="2025-11-27 17:11:53.314809402 +0000 UTC m=+148.587266754" lastFinishedPulling="2025-11-27 17:12:42.706036143 +0000 UTC m=+197.978493495" observedRunningTime="2025-11-27 17:12:43.129476687 +0000 UTC m=+198.401934039" watchObservedRunningTime="2025-11-27 17:12:43.130349352 +0000 UTC m=+198.402806694" Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.148248 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g6xz6" podStartSLOduration=2.608668511 podStartE2EDuration="55.148231048s" podCreationTimestamp="2025-11-27 17:11:48 +0000 UTC" firstStartedPulling="2025-11-27 17:11:50.137648274 +0000 UTC m=+145.410105616" lastFinishedPulling="2025-11-27 17:12:42.677210801 +0000 UTC m=+197.949668153" observedRunningTime="2025-11-27 17:12:43.147908878 +0000 UTC m=+198.420366220" watchObservedRunningTime="2025-11-27 17:12:43.148231048 +0000 UTC m=+198.420688400" Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.315544 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jsvqm"] Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.316034 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jsvqm" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerName="registry-server" containerID="cri-o://31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f" gracePeriod=2 Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.651833 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.784926 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-utilities\") pod \"824326f1-6377-4f05-a5fd-db94363ebe1d\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.784995 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nczqc\" (UniqueName: \"kubernetes.io/projected/824326f1-6377-4f05-a5fd-db94363ebe1d-kube-api-access-nczqc\") pod \"824326f1-6377-4f05-a5fd-db94363ebe1d\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.785093 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-catalog-content\") pod \"824326f1-6377-4f05-a5fd-db94363ebe1d\" (UID: \"824326f1-6377-4f05-a5fd-db94363ebe1d\") " Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.785701 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-utilities" (OuterVolumeSpecName: "utilities") pod "824326f1-6377-4f05-a5fd-db94363ebe1d" (UID: "824326f1-6377-4f05-a5fd-db94363ebe1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.792041 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/824326f1-6377-4f05-a5fd-db94363ebe1d-kube-api-access-nczqc" (OuterVolumeSpecName: "kube-api-access-nczqc") pod "824326f1-6377-4f05-a5fd-db94363ebe1d" (UID: "824326f1-6377-4f05-a5fd-db94363ebe1d"). InnerVolumeSpecName "kube-api-access-nczqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.886784 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.886822 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nczqc\" (UniqueName: \"kubernetes.io/projected/824326f1-6377-4f05-a5fd-db94363ebe1d-kube-api-access-nczqc\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.919337 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "824326f1-6377-4f05-a5fd-db94363ebe1d" (UID: "824326f1-6377-4f05-a5fd-db94363ebe1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:12:43 crc kubenswrapper[4809]: I1127 17:12:43.987781 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824326f1-6377-4f05-a5fd-db94363ebe1d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.075376 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrw24" event={"ID":"b1d9678d-171a-427b-9b37-192f523e63b2","Type":"ContainerStarted","Data":"528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d"} Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.084634 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hk4l" event={"ID":"03562c92-cba5-4c9a-b759-7da40962be62","Type":"ContainerStarted","Data":"a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493"} Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.089444 4809 generic.go:334] "Generic (PLEG): container finished" podID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerID="31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f" exitCode=0 Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.089485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsvqm" event={"ID":"824326f1-6377-4f05-a5fd-db94363ebe1d","Type":"ContainerDied","Data":"31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f"} Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.089512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jsvqm" event={"ID":"824326f1-6377-4f05-a5fd-db94363ebe1d","Type":"ContainerDied","Data":"30b810804d431d1b4daf51edfd5ebaa4814f064b198a307c7f34fb0ebce2d4d7"} Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.089516 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jsvqm" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.089535 4809 scope.go:117] "RemoveContainer" containerID="31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.114799 4809 scope.go:117] "RemoveContainer" containerID="0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.131837 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7hk4l" podStartSLOduration=1.5866907700000001 podStartE2EDuration="56.131814229s" podCreationTimestamp="2025-11-27 17:11:48 +0000 UTC" firstStartedPulling="2025-11-27 17:11:49.089003375 +0000 UTC m=+144.361460727" lastFinishedPulling="2025-11-27 17:12:43.634126844 +0000 UTC m=+198.906584186" observedRunningTime="2025-11-27 17:12:44.126038902 +0000 UTC m=+199.398496254" watchObservedRunningTime="2025-11-27 17:12:44.131814229 +0000 UTC m=+199.404271581" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.141550 4809 scope.go:117] "RemoveContainer" containerID="933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.165827 4809 scope.go:117] "RemoveContainer" containerID="31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.170334 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jsvqm"] Nov 27 17:12:44 crc kubenswrapper[4809]: E1127 17:12:44.171012 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f\": container with ID starting with 31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f not found: ID does not exist" containerID="31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.171065 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f"} err="failed to get container status \"31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f\": rpc error: code = NotFound desc = could not find container \"31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f\": container with ID starting with 31bbfb21290626b9e6b7eee9b7070c5a649dd1ea9214f2379fce72d1f988760f not found: ID does not exist" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.171132 4809 scope.go:117] "RemoveContainer" containerID="0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624" Nov 27 17:12:44 crc kubenswrapper[4809]: E1127 17:12:44.173297 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624\": container with ID starting with 0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624 not found: ID does not exist" containerID="0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.173357 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624"} err="failed to get container status \"0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624\": rpc error: code = NotFound desc = could not find container \"0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624\": container with ID starting with 0f2d3515524bb06ee338da5255dab93bb17304bb13c6579a3039d40e7347b624 not found: ID does not exist" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.173392 4809 scope.go:117] "RemoveContainer" containerID="933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554" Nov 27 17:12:44 crc kubenswrapper[4809]: E1127 17:12:44.173999 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554\": container with ID starting with 933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554 not found: ID does not exist" containerID="933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.174021 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554"} err="failed to get container status \"933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554\": rpc error: code = NotFound desc = could not find container \"933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554\": container with ID starting with 933fc48c7cae2c14ddb3bec7c2369f89445ff2de5d807a5370f1a7fdda936554 not found: ID does not exist" Nov 27 17:12:44 crc kubenswrapper[4809]: I1127 17:12:44.175417 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jsvqm"] Nov 27 17:12:45 crc kubenswrapper[4809]: I1127 17:12:45.097891 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcvzh" event={"ID":"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40","Type":"ContainerStarted","Data":"92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a"} Nov 27 17:12:45 crc kubenswrapper[4809]: I1127 17:12:45.100223 4809 generic.go:334] "Generic (PLEG): container finished" podID="b1d9678d-171a-427b-9b37-192f523e63b2" containerID="528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d" exitCode=0 Nov 27 17:12:45 crc kubenswrapper[4809]: I1127 17:12:45.100299 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrw24" event={"ID":"b1d9678d-171a-427b-9b37-192f523e63b2","Type":"ContainerDied","Data":"528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d"} Nov 27 17:12:45 crc kubenswrapper[4809]: I1127 17:12:45.105422 4809 generic.go:334] "Generic (PLEG): container finished" podID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerID="88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d" exitCode=0 Nov 27 17:12:45 crc kubenswrapper[4809]: I1127 17:12:45.105470 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbk" event={"ID":"72f7df98-e370-4a9b-bde4-f5ebccd29aca","Type":"ContainerDied","Data":"88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d"} Nov 27 17:12:45 crc kubenswrapper[4809]: I1127 17:12:45.123942 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dcvzh" podStartSLOduration=2.368184845 podStartE2EDuration="54.123922301s" podCreationTimestamp="2025-11-27 17:11:51 +0000 UTC" firstStartedPulling="2025-11-27 17:11:52.292926335 +0000 UTC m=+147.565383687" lastFinishedPulling="2025-11-27 17:12:44.048663791 +0000 UTC m=+199.321121143" observedRunningTime="2025-11-27 17:12:45.121066454 +0000 UTC m=+200.393523816" watchObservedRunningTime="2025-11-27 17:12:45.123922301 +0000 UTC m=+200.396379653" Nov 27 17:12:45 crc kubenswrapper[4809]: I1127 17:12:45.466971 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" path="/var/lib/kubelet/pods/824326f1-6377-4f05-a5fd-db94363ebe1d/volumes" Nov 27 17:12:48 crc kubenswrapper[4809]: I1127 17:12:48.394760 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:12:48 crc kubenswrapper[4809]: I1127 17:12:48.395167 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:12:48 crc kubenswrapper[4809]: I1127 17:12:48.440454 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:12:48 crc kubenswrapper[4809]: I1127 17:12:48.767110 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:12:48 crc kubenswrapper[4809]: I1127 17:12:48.767549 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:12:48 crc kubenswrapper[4809]: I1127 17:12:48.813265 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:12:49 crc kubenswrapper[4809]: I1127 17:12:49.170172 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:12:49 crc kubenswrapper[4809]: I1127 17:12:49.171180 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:12:49 crc kubenswrapper[4809]: I1127 17:12:49.710033 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g6xz6"] Nov 27 17:12:50 crc kubenswrapper[4809]: I1127 17:12:50.134396 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbk" event={"ID":"72f7df98-e370-4a9b-bde4-f5ebccd29aca","Type":"ContainerStarted","Data":"6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157"} Nov 27 17:12:50 crc kubenswrapper[4809]: I1127 17:12:50.399490 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:12:50 crc kubenswrapper[4809]: I1127 17:12:50.399555 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:12:50 crc kubenswrapper[4809]: I1127 17:12:50.451798 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:12:50 crc kubenswrapper[4809]: I1127 17:12:50.785051 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:12:50 crc kubenswrapper[4809]: I1127 17:12:50.785362 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:12:50 crc kubenswrapper[4809]: I1127 17:12:50.829087 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:12:51 crc kubenswrapper[4809]: I1127 17:12:51.152057 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrw24" event={"ID":"b1d9678d-171a-427b-9b37-192f523e63b2","Type":"ContainerStarted","Data":"923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd"} Nov 27 17:12:51 crc kubenswrapper[4809]: I1127 17:12:51.152108 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g6xz6" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerName="registry-server" containerID="cri-o://11bb3554f0667ce724c0695fb33793d3db6381f2f8da4acb48b39daec19b9c04" gracePeriod=2 Nov 27 17:12:51 crc kubenswrapper[4809]: I1127 17:12:51.182662 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-98fbk" podStartSLOduration=4.6661635740000005 podStartE2EDuration="1m3.182636782s" podCreationTimestamp="2025-11-27 17:11:48 +0000 UTC" firstStartedPulling="2025-11-27 17:11:50.116477134 +0000 UTC m=+145.388934476" lastFinishedPulling="2025-11-27 17:12:48.632950332 +0000 UTC m=+203.905407684" observedRunningTime="2025-11-27 17:12:51.179315932 +0000 UTC m=+206.451773324" watchObservedRunningTime="2025-11-27 17:12:51.182636782 +0000 UTC m=+206.455094134" Nov 27 17:12:51 crc kubenswrapper[4809]: I1127 17:12:51.204009 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:12:51 crc kubenswrapper[4809]: I1127 17:12:51.213660 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hrw24" podStartSLOduration=3.502201389 podStartE2EDuration="1m4.213624423s" podCreationTimestamp="2025-11-27 17:11:47 +0000 UTC" firstStartedPulling="2025-11-27 17:11:50.117120502 +0000 UTC m=+145.389577854" lastFinishedPulling="2025-11-27 17:12:50.828543536 +0000 UTC m=+206.101000888" observedRunningTime="2025-11-27 17:12:51.210223291 +0000 UTC m=+206.482680643" watchObservedRunningTime="2025-11-27 17:12:51.213624423 +0000 UTC m=+206.486081775" Nov 27 17:12:51 crc kubenswrapper[4809]: I1127 17:12:51.219057 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:12:51 crc kubenswrapper[4809]: I1127 17:12:51.391580 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:12:51 crc kubenswrapper[4809]: I1127 17:12:51.391655 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:12:51 crc kubenswrapper[4809]: I1127 17:12:51.433952 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.162690 4809 generic.go:334] "Generic (PLEG): container finished" podID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerID="11bb3554f0667ce724c0695fb33793d3db6381f2f8da4acb48b39daec19b9c04" exitCode=0 Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.162889 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g6xz6" event={"ID":"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83","Type":"ContainerDied","Data":"11bb3554f0667ce724c0695fb33793d3db6381f2f8da4acb48b39daec19b9c04"} Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.215138 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.217450 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.304830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnp4t\" (UniqueName: \"kubernetes.io/projected/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-kube-api-access-fnp4t\") pod \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.304894 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-utilities\") pod \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.305029 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-catalog-content\") pod \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\" (UID: \"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83\") " Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.305985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-utilities" (OuterVolumeSpecName: "utilities") pod "289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" (UID: "289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.311626 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.313981 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-kube-api-access-fnp4t" (OuterVolumeSpecName: "kube-api-access-fnp4t") pod "289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" (UID: "289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83"). InnerVolumeSpecName "kube-api-access-fnp4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.358481 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" (UID: "289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.412930 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnp4t\" (UniqueName: \"kubernetes.io/projected/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-kube-api-access-fnp4t\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:52 crc kubenswrapper[4809]: I1127 17:12:52.412980 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:53 crc kubenswrapper[4809]: I1127 17:12:53.172947 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g6xz6" Nov 27 17:12:53 crc kubenswrapper[4809]: I1127 17:12:53.173450 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g6xz6" event={"ID":"289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83","Type":"ContainerDied","Data":"73182e9db74c63d149e63f73f4252a87fbbeb1c3f45602c173fb8889cd1e6463"} Nov 27 17:12:53 crc kubenswrapper[4809]: I1127 17:12:53.173487 4809 scope.go:117] "RemoveContainer" containerID="11bb3554f0667ce724c0695fb33793d3db6381f2f8da4acb48b39daec19b9c04" Nov 27 17:12:53 crc kubenswrapper[4809]: I1127 17:12:53.190614 4809 scope.go:117] "RemoveContainer" containerID="27dff38a6eeee8d54ee2a38c30f447edae3c769c14a5ca5714965a1e327c2469" Nov 27 17:12:53 crc kubenswrapper[4809]: I1127 17:12:53.203881 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g6xz6"] Nov 27 17:12:53 crc kubenswrapper[4809]: I1127 17:12:53.211821 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g6xz6"] Nov 27 17:12:53 crc kubenswrapper[4809]: I1127 17:12:53.222905 4809 scope.go:117] "RemoveContainer" containerID="adc31ca2caefd2101822c3d9be2bc29282b164b3ea9ecb913ed6f44f97ad4cd9" Nov 27 17:12:53 crc kubenswrapper[4809]: I1127 17:12:53.464903 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" path="/var/lib/kubelet/pods/289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83/volumes" Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.309412 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrfjz"] Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.309616 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xrfjz" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerName="registry-server" containerID="cri-o://0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017" gracePeriod=2 Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.617067 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.745896 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-utilities\") pod \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.745952 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-catalog-content\") pod \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.746046 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llqvm\" (UniqueName: \"kubernetes.io/projected/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-kube-api-access-llqvm\") pod \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\" (UID: \"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7\") " Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.746728 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-utilities" (OuterVolumeSpecName: "utilities") pod "6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" (UID: "6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.752039 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-kube-api-access-llqvm" (OuterVolumeSpecName: "kube-api-access-llqvm") pod "6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" (UID: "6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7"). InnerVolumeSpecName "kube-api-access-llqvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.766176 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" (UID: "6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.847065 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llqvm\" (UniqueName: \"kubernetes.io/projected/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-kube-api-access-llqvm\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.847115 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:54 crc kubenswrapper[4809]: I1127 17:12:54.847125 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.185732 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerID="0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017" exitCode=0 Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.185798 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrfjz" event={"ID":"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7","Type":"ContainerDied","Data":"0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017"} Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.185827 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrfjz" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.185850 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrfjz" event={"ID":"6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7","Type":"ContainerDied","Data":"d6c4daee56ddfb0ec1d5f43c6be9665c1ec06559de780c21cd1019a36cacddaf"} Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.185868 4809 scope.go:117] "RemoveContainer" containerID="0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.206200 4809 scope.go:117] "RemoveContainer" containerID="1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.219619 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrfjz"] Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.224909 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrfjz"] Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.244727 4809 scope.go:117] "RemoveContainer" containerID="4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.261866 4809 scope.go:117] "RemoveContainer" containerID="0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017" Nov 27 17:12:55 crc kubenswrapper[4809]: E1127 17:12:55.262516 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017\": container with ID starting with 0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017 not found: ID does not exist" containerID="0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.262551 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017"} err="failed to get container status \"0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017\": rpc error: code = NotFound desc = could not find container \"0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017\": container with ID starting with 0e00ebb45e338a69bd55b974fd784f8a6497b0024fa78952a2c2677db0715017 not found: ID does not exist" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.262576 4809 scope.go:117] "RemoveContainer" containerID="1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243" Nov 27 17:12:55 crc kubenswrapper[4809]: E1127 17:12:55.262924 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243\": container with ID starting with 1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243 not found: ID does not exist" containerID="1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.262945 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243"} err="failed to get container status \"1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243\": rpc error: code = NotFound desc = could not find container \"1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243\": container with ID starting with 1d08ece6ff40cd61006e7036b3d518686e55e59d20d3bd08e99249e19400a243 not found: ID does not exist" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.262961 4809 scope.go:117] "RemoveContainer" containerID="4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4" Nov 27 17:12:55 crc kubenswrapper[4809]: E1127 17:12:55.263335 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4\": container with ID starting with 4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4 not found: ID does not exist" containerID="4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.263363 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4"} err="failed to get container status \"4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4\": rpc error: code = NotFound desc = could not find container \"4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4\": container with ID starting with 4a3b262d4a343a2856f9631f989e7654df1a58d0e1fbb8413d97dbf0fe621db4 not found: ID does not exist" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.469285 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" path="/var/lib/kubelet/pods/6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7/volumes" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.780172 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.780277 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.780354 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.781275 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:12:55 crc kubenswrapper[4809]: I1127 17:12:55.781357 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75" gracePeriod=600 Nov 27 17:12:56 crc kubenswrapper[4809]: I1127 17:12:56.194150 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75" exitCode=0 Nov 27 17:12:56 crc kubenswrapper[4809]: I1127 17:12:56.194254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75"} Nov 27 17:12:56 crc kubenswrapper[4809]: I1127 17:12:56.194613 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"bc8b858c869d6867bc225e685a98fd0cfd9e143945af5b8b25449257c7e1fd3a"} Nov 27 17:12:59 crc kubenswrapper[4809]: I1127 17:12:59.185688 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:12:59 crc kubenswrapper[4809]: I1127 17:12:59.186759 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:12:59 crc kubenswrapper[4809]: I1127 17:12:59.227245 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:12:59 crc kubenswrapper[4809]: I1127 17:12:59.269495 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:12:59 crc kubenswrapper[4809]: I1127 17:12:59.410418 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:12:59 crc kubenswrapper[4809]: I1127 17:12:59.410513 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:12:59 crc kubenswrapper[4809]: I1127 17:12:59.450031 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:13:00 crc kubenswrapper[4809]: I1127 17:13:00.268729 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:13:00 crc kubenswrapper[4809]: I1127 17:13:00.511801 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-98fbk"] Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.234701 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-98fbk" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerName="registry-server" containerID="cri-o://6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157" gracePeriod=2 Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.590686 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.649184 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-utilities\") pod \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.649291 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-catalog-content\") pod \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.649458 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgc92\" (UniqueName: \"kubernetes.io/projected/72f7df98-e370-4a9b-bde4-f5ebccd29aca-kube-api-access-wgc92\") pod \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\" (UID: \"72f7df98-e370-4a9b-bde4-f5ebccd29aca\") " Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.650325 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-utilities" (OuterVolumeSpecName: "utilities") pod "72f7df98-e370-4a9b-bde4-f5ebccd29aca" (UID: "72f7df98-e370-4a9b-bde4-f5ebccd29aca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.655107 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72f7df98-e370-4a9b-bde4-f5ebccd29aca-kube-api-access-wgc92" (OuterVolumeSpecName: "kube-api-access-wgc92") pod "72f7df98-e370-4a9b-bde4-f5ebccd29aca" (UID: "72f7df98-e370-4a9b-bde4-f5ebccd29aca"). InnerVolumeSpecName "kube-api-access-wgc92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.704549 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72f7df98-e370-4a9b-bde4-f5ebccd29aca" (UID: "72f7df98-e370-4a9b-bde4-f5ebccd29aca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.751536 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.751880 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72f7df98-e370-4a9b-bde4-f5ebccd29aca-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:02 crc kubenswrapper[4809]: I1127 17:13:02.751960 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgc92\" (UniqueName: \"kubernetes.io/projected/72f7df98-e370-4a9b-bde4-f5ebccd29aca-kube-api-access-wgc92\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.241232 4809 generic.go:334] "Generic (PLEG): container finished" podID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerID="6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157" exitCode=0 Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.241346 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98fbk" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.241369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbk" event={"ID":"72f7df98-e370-4a9b-bde4-f5ebccd29aca","Type":"ContainerDied","Data":"6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157"} Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.241639 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbk" event={"ID":"72f7df98-e370-4a9b-bde4-f5ebccd29aca","Type":"ContainerDied","Data":"71f95ff2b52e54a226171840d8e5a2bf0513296bb7a31c591cb3dde1ea36a165"} Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.241672 4809 scope.go:117] "RemoveContainer" containerID="6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.262048 4809 scope.go:117] "RemoveContainer" containerID="88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.267434 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-98fbk"] Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.270778 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-98fbk"] Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.298367 4809 scope.go:117] "RemoveContainer" containerID="0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.312983 4809 scope.go:117] "RemoveContainer" containerID="6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157" Nov 27 17:13:03 crc kubenswrapper[4809]: E1127 17:13:03.313480 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157\": container with ID starting with 6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157 not found: ID does not exist" containerID="6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.313533 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157"} err="failed to get container status \"6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157\": rpc error: code = NotFound desc = could not find container \"6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157\": container with ID starting with 6fc1c52fde6a65b034e7a0d6d3fcb6cacbbc085454e55f3b671e65e8a42c6157 not found: ID does not exist" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.313571 4809 scope.go:117] "RemoveContainer" containerID="88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d" Nov 27 17:13:03 crc kubenswrapper[4809]: E1127 17:13:03.314211 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d\": container with ID starting with 88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d not found: ID does not exist" containerID="88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.314249 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d"} err="failed to get container status \"88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d\": rpc error: code = NotFound desc = could not find container \"88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d\": container with ID starting with 88dc39487e472d5d99d905bb79b215202df51730c9f84db1345c666c3ec84b5d not found: ID does not exist" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.314276 4809 scope.go:117] "RemoveContainer" containerID="0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138" Nov 27 17:13:03 crc kubenswrapper[4809]: E1127 17:13:03.314561 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138\": container with ID starting with 0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138 not found: ID does not exist" containerID="0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.314610 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138"} err="failed to get container status \"0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138\": rpc error: code = NotFound desc = could not find container \"0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138\": container with ID starting with 0b2070e30c76ca4fa5f0101e020373076dac5c3c91fb2c41c77730880f8de138 not found: ID does not exist" Nov 27 17:13:03 crc kubenswrapper[4809]: I1127 17:13:03.464853 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" path="/var/lib/kubelet/pods/72f7df98-e370-4a9b-bde4-f5ebccd29aca/volumes" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.437483 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" podUID="fb3ec517-319b-4bd9-87b9-19ee21f410b6" containerName="oauth-openshift" containerID="cri-o://78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168" gracePeriod=15 Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.814980 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853410 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-66456c6bb-knj6b"] Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853612 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerName="extract-utilities" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853623 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerName="extract-utilities" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853632 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853638 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853646 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerName="extract-content" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853653 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerName="extract-content" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853660 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerName="extract-utilities" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853666 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerName="extract-utilities" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853685 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853698 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853709 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerName="extract-utilities" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853715 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerName="extract-utilities" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853723 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerName="extract-content" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853729 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerName="extract-content" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853765 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853774 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853781 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3ec517-319b-4bd9-87b9-19ee21f410b6" containerName="oauth-openshift" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853787 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3ec517-319b-4bd9-87b9-19ee21f410b6" containerName="oauth-openshift" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853793 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerName="extract-content" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853798 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerName="extract-content" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853807 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db95db91-9ee8-4b4b-b823-bb0b7a03a56c" containerName="pruner" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853812 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="db95db91-9ee8-4b4b-b823-bb0b7a03a56c" containerName="pruner" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853825 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerName="extract-utilities" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853831 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerName="extract-utilities" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853839 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853844 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: E1127 17:13:04.853854 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerName="extract-content" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853859 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerName="extract-content" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853949 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b2f80bd-419e-4d4a-b4b5-e6e7362a20a7" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853962 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="72f7df98-e370-4a9b-bde4-f5ebccd29aca" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853969 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="824326f1-6377-4f05-a5fd-db94363ebe1d" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853976 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="289e6fa9-7a46-4b2b-8ea3-1bc2bc0c4a83" containerName="registry-server" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853986 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3ec517-319b-4bd9-87b9-19ee21f410b6" containerName="oauth-openshift" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.853993 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="db95db91-9ee8-4b4b-b823-bb0b7a03a56c" containerName="pruner" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.854365 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.867496 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66456c6bb-knj6b"] Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.885073 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-session\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.885684 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-policies\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.885760 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-ocp-branding-template\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.885787 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-idp-0-file-data\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.885819 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-serving-cert\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.885874 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-login\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886086 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-trusted-ca-bundle\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886354 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886532 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-router-certs\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886572 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-cliconfig\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886604 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbb4h\" (UniqueName: \"kubernetes.io/projected/fb3ec517-319b-4bd9-87b9-19ee21f410b6-kube-api-access-qbb4h\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886641 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-service-ca\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886664 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-dir\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886679 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886702 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-provider-selection\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886730 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-error\") pod \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\" (UID: \"fb3ec517-319b-4bd9-87b9-19ee21f410b6\") " Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886797 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.886924 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.887197 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.887213 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.887226 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.887266 4809 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb3ec517-319b-4bd9-87b9-19ee21f410b6-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.887327 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.890534 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.891394 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb3ec517-319b-4bd9-87b9-19ee21f410b6-kube-api-access-qbb4h" (OuterVolumeSpecName: "kube-api-access-qbb4h") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "kube-api-access-qbb4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.891530 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.891668 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.891950 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.892242 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.896751 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.897280 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.897689 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "fb3ec517-319b-4bd9-87b9-19ee21f410b6" (UID: "fb3ec517-319b-4bd9-87b9-19ee21f410b6"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.987992 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988069 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-service-ca\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988105 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-router-certs\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-template-login\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988210 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-audit-policies\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988264 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988367 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988393 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988444 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-template-error\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvb4l\" (UniqueName: \"kubernetes.io/projected/07fa5b97-6983-4a4f-9e92-ad0b7d726339-kube-api-access-xvb4l\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988595 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/07fa5b97-6983-4a4f-9e92-ad0b7d726339-audit-dir\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988617 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-session\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988691 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988706 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988718 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988729 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988760 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988772 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988782 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988792 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988803 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb3ec517-319b-4bd9-87b9-19ee21f410b6-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:04 crc kubenswrapper[4809]: I1127 17:13:04.988813 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbb4h\" (UniqueName: \"kubernetes.io/projected/fb3ec517-319b-4bd9-87b9-19ee21f410b6-kube-api-access-qbb4h\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090431 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090461 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090480 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-template-error\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090503 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvb4l\" (UniqueName: \"kubernetes.io/projected/07fa5b97-6983-4a4f-9e92-ad0b7d726339-kube-api-access-xvb4l\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090538 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/07fa5b97-6983-4a4f-9e92-ad0b7d726339-audit-dir\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090554 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-session\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090574 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090606 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-service-ca\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090623 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-router-certs\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090639 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-template-login\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090657 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-audit-policies\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.090689 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.092929 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.092980 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-audit-policies\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.093154 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/07fa5b97-6983-4a4f-9e92-ad0b7d726339-audit-dir\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.094353 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.094803 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-template-error\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.094927 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-service-ca\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.094962 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-session\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.095130 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.095156 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-template-login\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.095276 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.095424 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.096012 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-router-certs\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.096023 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/07fa5b97-6983-4a4f-9e92-ad0b7d726339-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.108953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvb4l\" (UniqueName: \"kubernetes.io/projected/07fa5b97-6983-4a4f-9e92-ad0b7d726339-kube-api-access-xvb4l\") pod \"oauth-openshift-66456c6bb-knj6b\" (UID: \"07fa5b97-6983-4a4f-9e92-ad0b7d726339\") " pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.177429 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.268786 4809 generic.go:334] "Generic (PLEG): container finished" podID="fb3ec517-319b-4bd9-87b9-19ee21f410b6" containerID="78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168" exitCode=0 Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.268944 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" event={"ID":"fb3ec517-319b-4bd9-87b9-19ee21f410b6","Type":"ContainerDied","Data":"78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168"} Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.269147 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" event={"ID":"fb3ec517-319b-4bd9-87b9-19ee21f410b6","Type":"ContainerDied","Data":"e11d083fbf39eac27d07e2edb081c69071225d97bcdf6fc37f5fabfca6ed55df"} Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.269169 4809 scope.go:117] "RemoveContainer" containerID="78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.269017 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2t2fb" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.306723 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2t2fb"] Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.306914 4809 scope.go:117] "RemoveContainer" containerID="78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168" Nov 27 17:13:05 crc kubenswrapper[4809]: E1127 17:13:05.317541 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168\": container with ID starting with 78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168 not found: ID does not exist" containerID="78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.317582 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168"} err="failed to get container status \"78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168\": rpc error: code = NotFound desc = could not find container \"78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168\": container with ID starting with 78a2e996cbd8c9c5ffbb738e779c02bcb513093eb0dbe9df1128f75c77e6f168 not found: ID does not exist" Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.320787 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2t2fb"] Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.391114 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66456c6bb-knj6b"] Nov 27 17:13:05 crc kubenswrapper[4809]: W1127 17:13:05.399976 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07fa5b97_6983_4a4f_9e92_ad0b7d726339.slice/crio-6d0e6c87ffb217303549792c1474c70d6ae1d0646221e113f583d1df0c649f0b WatchSource:0}: Error finding container 6d0e6c87ffb217303549792c1474c70d6ae1d0646221e113f583d1df0c649f0b: Status 404 returned error can't find the container with id 6d0e6c87ffb217303549792c1474c70d6ae1d0646221e113f583d1df0c649f0b Nov 27 17:13:05 crc kubenswrapper[4809]: I1127 17:13:05.463764 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb3ec517-319b-4bd9-87b9-19ee21f410b6" path="/var/lib/kubelet/pods/fb3ec517-319b-4bd9-87b9-19ee21f410b6/volumes" Nov 27 17:13:06 crc kubenswrapper[4809]: I1127 17:13:06.282183 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" event={"ID":"07fa5b97-6983-4a4f-9e92-ad0b7d726339","Type":"ContainerStarted","Data":"c5dc090a6fed25f0eef2affc93ddf9b11ef63b04a2d929e5cb25dd169ab0a9b7"} Nov 27 17:13:06 crc kubenswrapper[4809]: I1127 17:13:06.282769 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:06 crc kubenswrapper[4809]: I1127 17:13:06.282800 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" event={"ID":"07fa5b97-6983-4a4f-9e92-ad0b7d726339","Type":"ContainerStarted","Data":"6d0e6c87ffb217303549792c1474c70d6ae1d0646221e113f583d1df0c649f0b"} Nov 27 17:13:06 crc kubenswrapper[4809]: I1127 17:13:06.287931 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" Nov 27 17:13:06 crc kubenswrapper[4809]: I1127 17:13:06.330170 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-66456c6bb-knj6b" podStartSLOduration=27.330147779 podStartE2EDuration="27.330147779s" podCreationTimestamp="2025-11-27 17:12:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:13:06.304876622 +0000 UTC m=+221.577333984" watchObservedRunningTime="2025-11-27 17:13:06.330147779 +0000 UTC m=+221.602605131" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.635967 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.636654 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.637493 4809 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.637805 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e" gracePeriod=15 Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.637861 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798" gracePeriod=15 Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.637894 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda" gracePeriod=15 Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.637827 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd" gracePeriod=15 Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.637822 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a" gracePeriod=15 Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639085 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 17:13:07 crc kubenswrapper[4809]: E1127 17:13:07.639430 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639449 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 17:13:07 crc kubenswrapper[4809]: E1127 17:13:07.639460 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639486 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 17:13:07 crc kubenswrapper[4809]: E1127 17:13:07.639497 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639503 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 17:13:07 crc kubenswrapper[4809]: E1127 17:13:07.639510 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639515 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 27 17:13:07 crc kubenswrapper[4809]: E1127 17:13:07.639526 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639531 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 17:13:07 crc kubenswrapper[4809]: E1127 17:13:07.639540 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639589 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 17:13:07 crc kubenswrapper[4809]: E1127 17:13:07.639604 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639611 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 17:13:07 crc kubenswrapper[4809]: E1127 17:13:07.639644 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639653 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639844 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639861 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639870 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639879 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639915 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.639925 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.640370 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.667751 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.725470 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.725597 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.725630 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.725702 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.725790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.725846 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.725909 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.725939 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827505 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827610 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827651 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827650 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827797 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827728 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827868 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827927 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827981 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827985 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.827999 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.828042 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.828091 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.828086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: I1127 17:13:07.965245 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:07 crc kubenswrapper[4809]: W1127 17:13:07.988104 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-cfc122141f3aa88ec8e24208d70d51f39c5970e038a9e54337630049d28d03e1 WatchSource:0}: Error finding container cfc122141f3aa88ec8e24208d70d51f39c5970e038a9e54337630049d28d03e1: Status 404 returned error can't find the container with id cfc122141f3aa88ec8e24208d70d51f39c5970e038a9e54337630049d28d03e1 Nov 27 17:13:07 crc kubenswrapper[4809]: E1127 17:13:07.991263 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.248:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187bec62c452ab71 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 17:13:07.990383473 +0000 UTC m=+223.262840826,LastTimestamp:2025-11-27 17:13:07.990383473 +0000 UTC m=+223.262840826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.298309 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.300232 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.301218 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a" exitCode=0 Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.301284 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798" exitCode=0 Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.301294 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd" exitCode=0 Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.301304 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda" exitCode=2 Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.301413 4809 scope.go:117] "RemoveContainer" containerID="13dc5c37e827c999defd09d12b80955a0a5b61963bab04bfb09588c55e1a3847" Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.307208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513"} Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.307285 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"cfc122141f3aa88ec8e24208d70d51f39c5970e038a9e54337630049d28d03e1"} Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.308081 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.308534 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.309635 4809 generic.go:334] "Generic (PLEG): container finished" podID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" containerID="467e6364808ccaab2258fb7ca72a66e4cc15eb93b844720a9319493d61baccaf" exitCode=0 Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.310468 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9","Type":"ContainerDied","Data":"467e6364808ccaab2258fb7ca72a66e4cc15eb93b844720a9319493d61baccaf"} Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.311325 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.312009 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:08 crc kubenswrapper[4809]: I1127 17:13:08.312717 4809 status_manager.go:851] "Failed to get status for pod" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.321357 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.607861 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.609295 4809 status_manager.go:851] "Failed to get status for pod" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.609817 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.754830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-var-lock\") pod \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.755327 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kubelet-dir\") pod \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.755398 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kube-api-access\") pod \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\" (UID: \"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9\") " Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.755868 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" (UID: "6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.755906 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-var-lock" (OuterVolumeSpecName: "var-lock") pod "6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" (UID: "6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.774996 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" (UID: "6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.856954 4809 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-var-lock\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.857018 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:09 crc kubenswrapper[4809]: I1127 17:13:09.857032 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.030557 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.032658 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.034151 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.034691 4809 status_manager.go:851] "Failed to get status for pod" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.035178 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.161671 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.161910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.161985 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.162032 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.162059 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.162135 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.162543 4809 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.162591 4809 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.162616 4809 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.331171 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9","Type":"ContainerDied","Data":"07005d6a3ef8d44153eb4b1b15242282faacad1d80e2c3702140177195b95559"} Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.331295 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.331265 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07005d6a3ef8d44153eb4b1b15242282faacad1d80e2c3702140177195b95559" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.336455 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.339080 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e" exitCode=0 Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.339167 4809 scope.go:117] "RemoveContainer" containerID="1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.339188 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.355571 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.355911 4809 status_manager.go:851] "Failed to get status for pod" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.357703 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.361477 4809 scope.go:117] "RemoveContainer" containerID="a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.363004 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.363600 4809 status_manager.go:851] "Failed to get status for pod" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.363981 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.376850 4809 scope.go:117] "RemoveContainer" containerID="8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.391500 4809 scope.go:117] "RemoveContainer" containerID="95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.409671 4809 scope.go:117] "RemoveContainer" containerID="2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.427214 4809 scope.go:117] "RemoveContainer" containerID="2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.464984 4809 scope.go:117] "RemoveContainer" containerID="1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a" Nov 27 17:13:10 crc kubenswrapper[4809]: E1127 17:13:10.465644 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\": container with ID starting with 1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a not found: ID does not exist" containerID="1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.465689 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a"} err="failed to get container status \"1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\": rpc error: code = NotFound desc = could not find container \"1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a\": container with ID starting with 1af1b8c2157fcda1d4b432c6044a896fc55220dcfccee5624c230ee46b8ebe2a not found: ID does not exist" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.465720 4809 scope.go:117] "RemoveContainer" containerID="a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798" Nov 27 17:13:10 crc kubenswrapper[4809]: E1127 17:13:10.466330 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\": container with ID starting with a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798 not found: ID does not exist" containerID="a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.466389 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798"} err="failed to get container status \"a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\": rpc error: code = NotFound desc = could not find container \"a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798\": container with ID starting with a6d326dd5b5ad4756135f9af77c69a0f855fbd215e27c794d2a494c9743fe798 not found: ID does not exist" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.466407 4809 scope.go:117] "RemoveContainer" containerID="8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd" Nov 27 17:13:10 crc kubenswrapper[4809]: E1127 17:13:10.466727 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\": container with ID starting with 8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd not found: ID does not exist" containerID="8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.466783 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd"} err="failed to get container status \"8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\": rpc error: code = NotFound desc = could not find container \"8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd\": container with ID starting with 8e8b096bb302bf7da17a4c2a95fefe6088d0d3a97b86566cac7e447b794e57dd not found: ID does not exist" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.466800 4809 scope.go:117] "RemoveContainer" containerID="95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda" Nov 27 17:13:10 crc kubenswrapper[4809]: E1127 17:13:10.467109 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\": container with ID starting with 95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda not found: ID does not exist" containerID="95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.467140 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda"} err="failed to get container status \"95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\": rpc error: code = NotFound desc = could not find container \"95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda\": container with ID starting with 95f0e385837b6a5423a975093eb997e14d396adee4e4aaf773ae3995d7496fda not found: ID does not exist" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.467157 4809 scope.go:117] "RemoveContainer" containerID="2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e" Nov 27 17:13:10 crc kubenswrapper[4809]: E1127 17:13:10.467423 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\": container with ID starting with 2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e not found: ID does not exist" containerID="2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.467477 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e"} err="failed to get container status \"2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\": rpc error: code = NotFound desc = could not find container \"2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e\": container with ID starting with 2d3df8b6d80311b4f28f2c411ca3f5e004dba676bb6ab186e85bec412e7a7c2e not found: ID does not exist" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.467492 4809 scope.go:117] "RemoveContainer" containerID="2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59" Nov 27 17:13:10 crc kubenswrapper[4809]: E1127 17:13:10.467814 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\": container with ID starting with 2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59 not found: ID does not exist" containerID="2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59" Nov 27 17:13:10 crc kubenswrapper[4809]: I1127 17:13:10.467845 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59"} err="failed to get container status \"2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\": rpc error: code = NotFound desc = could not find container \"2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59\": container with ID starting with 2cca82ce2ebe87825c1bc79198f4fe4cfc6a5cae6d135db2eaf4e1608e541f59 not found: ID does not exist" Nov 27 17:13:11 crc kubenswrapper[4809]: I1127 17:13:11.467106 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 27 17:13:13 crc kubenswrapper[4809]: E1127 17:13:13.803119 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.248:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187bec62c452ab71 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 17:13:07.990383473 +0000 UTC m=+223.262840826,LastTimestamp:2025-11-27 17:13:07.990383473 +0000 UTC m=+223.262840826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 17:13:15 crc kubenswrapper[4809]: E1127 17:13:15.093189 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:15 crc kubenswrapper[4809]: E1127 17:13:15.093987 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:15 crc kubenswrapper[4809]: E1127 17:13:15.094890 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:15 crc kubenswrapper[4809]: E1127 17:13:15.095259 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:15 crc kubenswrapper[4809]: E1127 17:13:15.095637 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:15 crc kubenswrapper[4809]: I1127 17:13:15.095678 4809 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 27 17:13:15 crc kubenswrapper[4809]: E1127 17:13:15.095907 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="200ms" Nov 27 17:13:15 crc kubenswrapper[4809]: E1127 17:13:15.297733 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="400ms" Nov 27 17:13:15 crc kubenswrapper[4809]: I1127 17:13:15.465409 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:15 crc kubenswrapper[4809]: I1127 17:13:15.465917 4809 status_manager.go:851] "Failed to get status for pod" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:15 crc kubenswrapper[4809]: E1127 17:13:15.698939 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="800ms" Nov 27 17:13:16 crc kubenswrapper[4809]: E1127 17:13:16.495673 4809 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.248:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" volumeName="registry-storage" Nov 27 17:13:16 crc kubenswrapper[4809]: E1127 17:13:16.499932 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="1.6s" Nov 27 17:13:18 crc kubenswrapper[4809]: E1127 17:13:18.101482 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" interval="3.2s" Nov 27 17:13:19 crc kubenswrapper[4809]: I1127 17:13:19.460394 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:19 crc kubenswrapper[4809]: I1127 17:13:19.461987 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:19 crc kubenswrapper[4809]: I1127 17:13:19.462618 4809 status_manager.go:851] "Failed to get status for pod" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:19 crc kubenswrapper[4809]: I1127 17:13:19.477249 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:19 crc kubenswrapper[4809]: I1127 17:13:19.477284 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:19 crc kubenswrapper[4809]: E1127 17:13:19.477870 4809 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:19 crc kubenswrapper[4809]: I1127 17:13:19.478872 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:19 crc kubenswrapper[4809]: E1127 17:13:19.983394 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:13:19Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:13:19Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:13:19Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T17:13:19Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:19 crc kubenswrapper[4809]: E1127 17:13:19.984068 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:19 crc kubenswrapper[4809]: E1127 17:13:19.984474 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:19 crc kubenswrapper[4809]: E1127 17:13:19.984960 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:19 crc kubenswrapper[4809]: E1127 17:13:19.985185 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:19 crc kubenswrapper[4809]: E1127 17:13:19.985207 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 17:13:20 crc kubenswrapper[4809]: I1127 17:13:20.408861 4809 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b610d97591eaa36f6811509a39dd1cb15642b006ea581f5458aac487e2dfdb3d" exitCode=0 Nov 27 17:13:20 crc kubenswrapper[4809]: I1127 17:13:20.409075 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b610d97591eaa36f6811509a39dd1cb15642b006ea581f5458aac487e2dfdb3d"} Nov 27 17:13:20 crc kubenswrapper[4809]: I1127 17:13:20.409443 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0a9c2a0c00f8327de1c38f9164f8d12bf3d7f9fae8f3c0d7f389a51d61813130"} Nov 27 17:13:20 crc kubenswrapper[4809]: I1127 17:13:20.409907 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:20 crc kubenswrapper[4809]: I1127 17:13:20.409928 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:20 crc kubenswrapper[4809]: E1127 17:13:20.410597 4809 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:20 crc kubenswrapper[4809]: I1127 17:13:20.410580 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:20 crc kubenswrapper[4809]: I1127 17:13:20.411478 4809 status_manager.go:851] "Failed to get status for pod" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.248:6443: connect: connection refused" Nov 27 17:13:21 crc kubenswrapper[4809]: I1127 17:13:21.150064 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 27 17:13:21 crc kubenswrapper[4809]: I1127 17:13:21.150488 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 27 17:13:21 crc kubenswrapper[4809]: I1127 17:13:21.435616 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ac8973b650fdc665cd0ad661baa71b694bfb506d846d373ff033896e6242b6eb"} Nov 27 17:13:21 crc kubenswrapper[4809]: I1127 17:13:21.435680 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"38b97a4504fad3e5f197b82c9c62e64f18722c330f8ee7b2854e6a6d381283a2"} Nov 27 17:13:21 crc kubenswrapper[4809]: I1127 17:13:21.435691 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"367030e2f88539c2a95caa80424facbc13bd33a8bdcc6dd0d81abf5276d88f3d"} Nov 27 17:13:21 crc kubenswrapper[4809]: I1127 17:13:21.450818 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 27 17:13:21 crc kubenswrapper[4809]: I1127 17:13:21.450899 4809 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8" exitCode=1 Nov 27 17:13:21 crc kubenswrapper[4809]: I1127 17:13:21.450953 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8"} Nov 27 17:13:21 crc kubenswrapper[4809]: I1127 17:13:21.451781 4809 scope.go:117] "RemoveContainer" containerID="501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8" Nov 27 17:13:22 crc kubenswrapper[4809]: I1127 17:13:22.127257 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:13:22 crc kubenswrapper[4809]: I1127 17:13:22.461800 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 27 17:13:22 crc kubenswrapper[4809]: I1127 17:13:22.461990 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"19fa97fbea3a99107b7fa42ee61805ff9ae76a7c8148a1a4cdf07150088eea71"} Nov 27 17:13:22 crc kubenswrapper[4809]: I1127 17:13:22.466871 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d5c23e7260a1ad0068f71b1188d9e47d61f5c12a6e30d1c7bc17e936af00822a"} Nov 27 17:13:22 crc kubenswrapper[4809]: I1127 17:13:22.466943 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c106591e0648c0180cdf8ba74e0b6cb3e5aff3d91a438cdc4fa21f26671508dc"} Nov 27 17:13:22 crc kubenswrapper[4809]: I1127 17:13:22.467197 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:22 crc kubenswrapper[4809]: I1127 17:13:22.467353 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:22 crc kubenswrapper[4809]: I1127 17:13:22.467390 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:24 crc kubenswrapper[4809]: I1127 17:13:24.479312 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:24 crc kubenswrapper[4809]: I1127 17:13:24.479693 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:24 crc kubenswrapper[4809]: I1127 17:13:24.486312 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:27 crc kubenswrapper[4809]: I1127 17:13:27.475818 4809 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:27 crc kubenswrapper[4809]: I1127 17:13:27.492168 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:27 crc kubenswrapper[4809]: I1127 17:13:27.492196 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:27 crc kubenswrapper[4809]: I1127 17:13:27.496322 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:27 crc kubenswrapper[4809]: I1127 17:13:27.498351 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="fd1df817-3d32-41e7-baf6-f03f90f8c178" Nov 27 17:13:28 crc kubenswrapper[4809]: I1127 17:13:28.497306 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:28 crc kubenswrapper[4809]: I1127 17:13:28.497349 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="518cb4a7-b48a-42fc-9436-13b6e54f9697" Nov 27 17:13:31 crc kubenswrapper[4809]: I1127 17:13:31.150059 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:13:32 crc kubenswrapper[4809]: I1127 17:13:32.126836 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:13:32 crc kubenswrapper[4809]: I1127 17:13:32.127352 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 27 17:13:32 crc kubenswrapper[4809]: I1127 17:13:32.127543 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 27 17:13:35 crc kubenswrapper[4809]: I1127 17:13:35.471548 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="fd1df817-3d32-41e7-baf6-f03f90f8c178" Nov 27 17:13:37 crc kubenswrapper[4809]: I1127 17:13:37.338900 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 27 17:13:38 crc kubenswrapper[4809]: I1127 17:13:38.344818 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 27 17:13:38 crc kubenswrapper[4809]: I1127 17:13:38.538212 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 27 17:13:38 crc kubenswrapper[4809]: I1127 17:13:38.592555 4809 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 27 17:13:38 crc kubenswrapper[4809]: I1127 17:13:38.608970 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 27 17:13:38 crc kubenswrapper[4809]: I1127 17:13:38.620019 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 27 17:13:38 crc kubenswrapper[4809]: I1127 17:13:38.859154 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 17:13:38 crc kubenswrapper[4809]: I1127 17:13:38.973610 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 27 17:13:38 crc kubenswrapper[4809]: I1127 17:13:38.994533 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 27 17:13:39 crc kubenswrapper[4809]: I1127 17:13:39.143992 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 27 17:13:39 crc kubenswrapper[4809]: I1127 17:13:39.352366 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 27 17:13:39 crc kubenswrapper[4809]: I1127 17:13:39.384455 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 27 17:13:39 crc kubenswrapper[4809]: I1127 17:13:39.831848 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 27 17:13:39 crc kubenswrapper[4809]: I1127 17:13:39.956682 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 27 17:13:39 crc kubenswrapper[4809]: I1127 17:13:39.987033 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.174944 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.242054 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.313913 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.324609 4809 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.440636 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.496287 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.549337 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.561242 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.684007 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.690275 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.759079 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.763678 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 17:13:40 crc kubenswrapper[4809]: I1127 17:13:40.805889 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.030952 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.098584 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.160625 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.222471 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.250875 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.251887 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.330461 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.341467 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.402805 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.412917 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.443092 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.451598 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.604805 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.727367 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.746162 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.796648 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 27 17:13:41 crc kubenswrapper[4809]: I1127 17:13:41.852701 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.092436 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.127637 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.127732 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.264326 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.338844 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.353397 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.396695 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.536587 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.566470 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.692579 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.709503 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.895470 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 17:13:42 crc kubenswrapper[4809]: I1127 17:13:42.981277 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.014544 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.018038 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.078545 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.258510 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.321634 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.343802 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.380720 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.415731 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.437318 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.456352 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.576096 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.603490 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.672916 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.691864 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.728594 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.772884 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.915800 4809 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.930997 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 27 17:13:43 crc kubenswrapper[4809]: I1127 17:13:43.964272 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.051553 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.148700 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.175854 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.257987 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.260005 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.266086 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.329652 4809 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.331020 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=37.330996708 podStartE2EDuration="37.330996708s" podCreationTimestamp="2025-11-27 17:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:13:27.233576878 +0000 UTC m=+242.506034240" watchObservedRunningTime="2025-11-27 17:13:44.330996708 +0000 UTC m=+259.603454060" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.334491 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.334540 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.340412 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.351985 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.359203 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.359181144 podStartE2EDuration="17.359181144s" podCreationTimestamp="2025-11-27 17:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:13:44.355556164 +0000 UTC m=+259.628013536" watchObservedRunningTime="2025-11-27 17:13:44.359181144 +0000 UTC m=+259.631638516" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.383491 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.582218 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.649983 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.798482 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.841909 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.857037 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 17:13:44 crc kubenswrapper[4809]: I1127 17:13:44.867898 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.137917 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.175263 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.239779 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.259102 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.266916 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.309725 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.352461 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.400492 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.471632 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.491286 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.619010 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.689615 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.747965 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.853364 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.855372 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.871364 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.895536 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.941531 4809 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 27 17:13:45 crc kubenswrapper[4809]: I1127 17:13:45.946677 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.031125 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.158339 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.165724 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.252144 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.314983 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.351503 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.368011 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.515954 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.545430 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.563413 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.574125 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.598640 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.615904 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.706503 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.708623 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.722179 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.728144 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.761055 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.764414 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.825761 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 27 17:13:46 crc kubenswrapper[4809]: I1127 17:13:46.862442 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.007681 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.021947 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.040017 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.082354 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.113481 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.138272 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.153700 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.228090 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.336238 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.349390 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.370671 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.407559 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.550198 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.628066 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.688595 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.811682 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.849270 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 27 17:13:47 crc kubenswrapper[4809]: I1127 17:13:47.907934 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.010487 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.134455 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.137044 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.234018 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.259780 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.357759 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.368636 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.439781 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.448525 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.464590 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.478283 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.502959 4809 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.524058 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.558108 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.673158 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.778598 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.802375 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 17:13:48 crc kubenswrapper[4809]: I1127 17:13:48.907783 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.027158 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.075525 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.195875 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.195894 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.253675 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.268700 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.287587 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.311717 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.348418 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.403319 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.412603 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.439677 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.502287 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.517936 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.561911 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.579809 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.626057 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.777915 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.831105 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.847232 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.867102 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.891459 4809 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.891681 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513" gracePeriod=5 Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.892870 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 27 17:13:49 crc kubenswrapper[4809]: I1127 17:13:49.956342 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.055102 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.145985 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.159611 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.160904 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.203507 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.301231 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.307999 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.312543 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.330527 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.349568 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.378250 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.400699 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.455675 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.557566 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.720042 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.764955 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.857004 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.868694 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 27 17:13:50 crc kubenswrapper[4809]: I1127 17:13:50.949893 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.026930 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.125809 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.208416 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.232426 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.249628 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.271620 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.374988 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.436872 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.683321 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 17:13:51 crc kubenswrapper[4809]: I1127 17:13:51.876658 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.127391 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.128435 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.128630 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.129787 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"19fa97fbea3a99107b7fa42ee61805ff9ae76a7c8148a1a4cdf07150088eea71"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.130139 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://19fa97fbea3a99107b7fa42ee61805ff9ae76a7c8148a1a4cdf07150088eea71" gracePeriod=30 Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.147096 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.252456 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.254246 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.282362 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.547965 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.645800 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 27 17:13:52 crc kubenswrapper[4809]: I1127 17:13:52.922243 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.048563 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.105667 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.247912 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.254522 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.466895 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.468354 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.469785 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.487026 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.559021 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.584870 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.599284 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.681644 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.778552 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.830234 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.900098 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 27 17:13:53 crc kubenswrapper[4809]: I1127 17:13:53.995187 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 27 17:13:54 crc kubenswrapper[4809]: I1127 17:13:54.268266 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 27 17:13:54 crc kubenswrapper[4809]: I1127 17:13:54.305215 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 27 17:13:54 crc kubenswrapper[4809]: I1127 17:13:54.468967 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 17:13:54 crc kubenswrapper[4809]: I1127 17:13:54.731369 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 17:13:54 crc kubenswrapper[4809]: I1127 17:13:54.774666 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 27 17:13:54 crc kubenswrapper[4809]: I1127 17:13:54.827572 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 27 17:13:54 crc kubenswrapper[4809]: I1127 17:13:54.957343 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.060620 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.461020 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.461097 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.463920 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.474122 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.474155 4809 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="e7883c8e-07ad-41fc-9b63-adc81fbb08bd" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.478372 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.478411 4809 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="e7883c8e-07ad-41fc-9b63-adc81fbb08bd" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.585473 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.627372 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.627461 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.627566 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.627635 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.627702 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.627772 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.627841 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.627880 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.627993 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.629135 4809 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.629173 4809 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.629200 4809 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.629214 4809 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.635947 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.654073 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.654152 4809 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513" exitCode=137 Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.654281 4809 scope.go:117] "RemoveContainer" containerID="510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.654343 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.681410 4809 scope.go:117] "RemoveContainer" containerID="510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513" Nov 27 17:13:55 crc kubenswrapper[4809]: E1127 17:13:55.681775 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513\": container with ID starting with 510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513 not found: ID does not exist" containerID="510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.681830 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513"} err="failed to get container status \"510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513\": rpc error: code = NotFound desc = could not find container \"510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513\": container with ID starting with 510d750589257a76098e37c446ef4adbddf745f560d2ed025df5657ff37d3513 not found: ID does not exist" Nov 27 17:13:55 crc kubenswrapper[4809]: I1127 17:13:55.730813 4809 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 17:13:56 crc kubenswrapper[4809]: I1127 17:13:56.090590 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 17:13:57 crc kubenswrapper[4809]: I1127 17:13:57.118280 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 27 17:13:57 crc kubenswrapper[4809]: I1127 17:13:57.464558 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 27 17:14:21 crc kubenswrapper[4809]: I1127 17:14:21.985671 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2cphh"] Nov 27 17:14:21 crc kubenswrapper[4809]: I1127 17:14:21.986505 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" podUID="23af03b4-a427-4e08-848b-0561b9cc7364" containerName="controller-manager" containerID="cri-o://970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0" gracePeriod=30 Nov 27 17:14:21 crc kubenswrapper[4809]: I1127 17:14:21.991579 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9"] Nov 27 17:14:21 crc kubenswrapper[4809]: I1127 17:14:21.991937 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" podUID="f6349461-b5ae-4e1a-9711-9d676f1ec01f" containerName="route-controller-manager" containerID="cri-o://f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217" gracePeriod=30 Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.346345 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.377394 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.380023 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6349461-b5ae-4e1a-9711-9d676f1ec01f-serving-cert\") pod \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.380066 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-proxy-ca-bundles\") pod \"23af03b4-a427-4e08-848b-0561b9cc7364\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.380118 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-config\") pod \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.380135 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-client-ca\") pod \"23af03b4-a427-4e08-848b-0561b9cc7364\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.380168 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b5rd\" (UniqueName: \"kubernetes.io/projected/23af03b4-a427-4e08-848b-0561b9cc7364-kube-api-access-9b5rd\") pod \"23af03b4-a427-4e08-848b-0561b9cc7364\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.380200 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23af03b4-a427-4e08-848b-0561b9cc7364-serving-cert\") pod \"23af03b4-a427-4e08-848b-0561b9cc7364\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.380220 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwvlz\" (UniqueName: \"kubernetes.io/projected/f6349461-b5ae-4e1a-9711-9d676f1ec01f-kube-api-access-pwvlz\") pod \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.380241 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-config\") pod \"23af03b4-a427-4e08-848b-0561b9cc7364\" (UID: \"23af03b4-a427-4e08-848b-0561b9cc7364\") " Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.381117 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "23af03b4-a427-4e08-848b-0561b9cc7364" (UID: "23af03b4-a427-4e08-848b-0561b9cc7364"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.381421 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-config" (OuterVolumeSpecName: "config") pod "f6349461-b5ae-4e1a-9711-9d676f1ec01f" (UID: "f6349461-b5ae-4e1a-9711-9d676f1ec01f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.381716 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-client-ca" (OuterVolumeSpecName: "client-ca") pod "23af03b4-a427-4e08-848b-0561b9cc7364" (UID: "23af03b4-a427-4e08-848b-0561b9cc7364"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.382223 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-config" (OuterVolumeSpecName: "config") pod "23af03b4-a427-4e08-848b-0561b9cc7364" (UID: "23af03b4-a427-4e08-848b-0561b9cc7364"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.385805 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6349461-b5ae-4e1a-9711-9d676f1ec01f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f6349461-b5ae-4e1a-9711-9d676f1ec01f" (UID: "f6349461-b5ae-4e1a-9711-9d676f1ec01f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.385884 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23af03b4-a427-4e08-848b-0561b9cc7364-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "23af03b4-a427-4e08-848b-0561b9cc7364" (UID: "23af03b4-a427-4e08-848b-0561b9cc7364"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.385997 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23af03b4-a427-4e08-848b-0561b9cc7364-kube-api-access-9b5rd" (OuterVolumeSpecName: "kube-api-access-9b5rd") pod "23af03b4-a427-4e08-848b-0561b9cc7364" (UID: "23af03b4-a427-4e08-848b-0561b9cc7364"). InnerVolumeSpecName "kube-api-access-9b5rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.386012 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6349461-b5ae-4e1a-9711-9d676f1ec01f-kube-api-access-pwvlz" (OuterVolumeSpecName: "kube-api-access-pwvlz") pod "f6349461-b5ae-4e1a-9711-9d676f1ec01f" (UID: "f6349461-b5ae-4e1a-9711-9d676f1ec01f"). InnerVolumeSpecName "kube-api-access-pwvlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.481012 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-client-ca\") pod \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\" (UID: \"f6349461-b5ae-4e1a-9711-9d676f1ec01f\") " Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.481534 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-client-ca" (OuterVolumeSpecName: "client-ca") pod "f6349461-b5ae-4e1a-9711-9d676f1ec01f" (UID: "f6349461-b5ae-4e1a-9711-9d676f1ec01f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.481998 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.482025 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.482037 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b5rd\" (UniqueName: \"kubernetes.io/projected/23af03b4-a427-4e08-848b-0561b9cc7364-kube-api-access-9b5rd\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.482048 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23af03b4-a427-4e08-848b-0561b9cc7364-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.482057 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwvlz\" (UniqueName: \"kubernetes.io/projected/f6349461-b5ae-4e1a-9711-9d676f1ec01f-kube-api-access-pwvlz\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.482066 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.482074 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6349461-b5ae-4e1a-9711-9d676f1ec01f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.482082 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23af03b4-a427-4e08-848b-0561b9cc7364-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.582771 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6349461-b5ae-4e1a-9711-9d676f1ec01f-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.818710 4809 generic.go:334] "Generic (PLEG): container finished" podID="f6349461-b5ae-4e1a-9711-9d676f1ec01f" containerID="f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217" exitCode=0 Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.818808 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" event={"ID":"f6349461-b5ae-4e1a-9711-9d676f1ec01f","Type":"ContainerDied","Data":"f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217"} Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.818834 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.818852 4809 scope.go:117] "RemoveContainer" containerID="f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.818839 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9" event={"ID":"f6349461-b5ae-4e1a-9711-9d676f1ec01f","Type":"ContainerDied","Data":"03480f478735a3b2d35fe52cf776d0b56ed99ece8a9560c9e06ff047a4ace28e"} Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.822399 4809 generic.go:334] "Generic (PLEG): container finished" podID="23af03b4-a427-4e08-848b-0561b9cc7364" containerID="970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0" exitCode=0 Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.822490 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.822535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" event={"ID":"23af03b4-a427-4e08-848b-0561b9cc7364","Type":"ContainerDied","Data":"970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0"} Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.822585 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2cphh" event={"ID":"23af03b4-a427-4e08-848b-0561b9cc7364","Type":"ContainerDied","Data":"dac87afede946ab918c20020e54ac5680187e2ff813db1db1e915627022993e1"} Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.824983 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.827296 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.827334 4809 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="19fa97fbea3a99107b7fa42ee61805ff9ae76a7c8148a1a4cdf07150088eea71" exitCode=137 Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.827407 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"19fa97fbea3a99107b7fa42ee61805ff9ae76a7c8148a1a4cdf07150088eea71"} Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.833456 4809 scope.go:117] "RemoveContainer" containerID="f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217" Nov 27 17:14:22 crc kubenswrapper[4809]: E1127 17:14:22.833923 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217\": container with ID starting with f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217 not found: ID does not exist" containerID="f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.833952 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217"} err="failed to get container status \"f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217\": rpc error: code = NotFound desc = could not find container \"f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217\": container with ID starting with f2a430ade7d2132f50eaadea2fdec21f740d9c5e8b109ae18c204d60e9ebd217 not found: ID does not exist" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.833973 4809 scope.go:117] "RemoveContainer" containerID="970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.847004 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9"] Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.852755 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wq2f9"] Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.853318 4809 scope.go:117] "RemoveContainer" containerID="970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0" Nov 27 17:14:22 crc kubenswrapper[4809]: E1127 17:14:22.853620 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0\": container with ID starting with 970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0 not found: ID does not exist" containerID="970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.853648 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0"} err="failed to get container status \"970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0\": rpc error: code = NotFound desc = could not find container \"970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0\": container with ID starting with 970d3922e48b2a1f73802378cadf02e3381d591a68bd63dd7962958f7881f9f0 not found: ID does not exist" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.853667 4809 scope.go:117] "RemoveContainer" containerID="501fcac85367757661114776efcee64a9d02f65ed03b8ab7752b79bdc7aaacd8" Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.855838 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2cphh"] Nov 27 17:14:22 crc kubenswrapper[4809]: I1127 17:14:22.858826 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2cphh"] Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.463880 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23af03b4-a427-4e08-848b-0561b9cc7364" path="/var/lib/kubelet/pods/23af03b4-a427-4e08-848b-0561b9cc7364/volumes" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.464415 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6349461-b5ae-4e1a-9711-9d676f1ec01f" path="/var/lib/kubelet/pods/f6349461-b5ae-4e1a-9711-9d676f1ec01f/volumes" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.833905 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.835025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a0ae18f8ed1e9a72e931d77e207780b9eaef95c9ea637190e7de2b8ac958e3cd"} Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.905470 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7bf57cbd5-ksszv"] Nov 27 17:14:23 crc kubenswrapper[4809]: E1127 17:14:23.905717 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.905746 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 17:14:23 crc kubenswrapper[4809]: E1127 17:14:23.905758 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" containerName="installer" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.905764 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" containerName="installer" Nov 27 17:14:23 crc kubenswrapper[4809]: E1127 17:14:23.905777 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23af03b4-a427-4e08-848b-0561b9cc7364" containerName="controller-manager" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.905785 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="23af03b4-a427-4e08-848b-0561b9cc7364" containerName="controller-manager" Nov 27 17:14:23 crc kubenswrapper[4809]: E1127 17:14:23.905799 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6349461-b5ae-4e1a-9711-9d676f1ec01f" containerName="route-controller-manager" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.905804 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6349461-b5ae-4e1a-9711-9d676f1ec01f" containerName="route-controller-manager" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.905889 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="23af03b4-a427-4e08-848b-0561b9cc7364" containerName="controller-manager" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.905899 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.905907 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d7fbe8a-82e8-421a-a2fd-8e2737f69bb9" containerName="installer" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.905916 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6349461-b5ae-4e1a-9711-9d676f1ec01f" containerName="route-controller-manager" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.906267 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.908766 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.908817 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.909172 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.909420 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68665d969-sctbd"] Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.910048 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.911172 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.911334 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.911364 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.912529 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.914234 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.914311 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.916282 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.916625 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.916836 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.918229 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.921848 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bf57cbd5-ksszv"] Nov 27 17:14:23 crc kubenswrapper[4809]: I1127 17:14:23.928074 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68665d969-sctbd"] Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.100243 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-serving-cert\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.100707 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14f4be19-469c-4ae1-af53-da001a4d6d33-serving-cert\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.100883 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-config\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.101018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-client-ca\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.101129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-client-ca\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.101285 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27rw9\" (UniqueName: \"kubernetes.io/projected/14f4be19-469c-4ae1-af53-da001a4d6d33-kube-api-access-27rw9\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.101473 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-config\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.101583 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-proxy-ca-bundles\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.101695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl2w7\" (UniqueName: \"kubernetes.io/projected/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-kube-api-access-hl2w7\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.202303 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-serving-cert\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.202349 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14f4be19-469c-4ae1-af53-da001a4d6d33-serving-cert\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.202370 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-config\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.202401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-client-ca\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.202420 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-client-ca\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.202439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27rw9\" (UniqueName: \"kubernetes.io/projected/14f4be19-469c-4ae1-af53-da001a4d6d33-kube-api-access-27rw9\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.202466 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-config\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.202481 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-proxy-ca-bundles\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.202499 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl2w7\" (UniqueName: \"kubernetes.io/projected/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-kube-api-access-hl2w7\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.206205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-proxy-ca-bundles\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.207153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-client-ca\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.207586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-config\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.207887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14f4be19-469c-4ae1-af53-da001a4d6d33-serving-cert\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.208280 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-serving-cert\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.208778 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-config\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.215211 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-client-ca\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.220012 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27rw9\" (UniqueName: \"kubernetes.io/projected/14f4be19-469c-4ae1-af53-da001a4d6d33-kube-api-access-27rw9\") pod \"route-controller-manager-68665d969-sctbd\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.221138 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl2w7\" (UniqueName: \"kubernetes.io/projected/25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae-kube-api-access-hl2w7\") pod \"controller-manager-7bf57cbd5-ksszv\" (UID: \"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae\") " pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.226011 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.232939 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.417492 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bf57cbd5-ksszv"] Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.443846 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68665d969-sctbd"] Nov 27 17:14:24 crc kubenswrapper[4809]: W1127 17:14:24.447704 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14f4be19_469c_4ae1_af53_da001a4d6d33.slice/crio-1fc8d3de5b4094c39efd379765a040ddf46ff276463e921cad0d3757c75b7f2c WatchSource:0}: Error finding container 1fc8d3de5b4094c39efd379765a040ddf46ff276463e921cad0d3757c75b7f2c: Status 404 returned error can't find the container with id 1fc8d3de5b4094c39efd379765a040ddf46ff276463e921cad0d3757c75b7f2c Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.844476 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" event={"ID":"14f4be19-469c-4ae1-af53-da001a4d6d33","Type":"ContainerStarted","Data":"04f26916c0840859c8c2594175b57ab2ea0a80ea4325f959f8ef8a570a90c1eb"} Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.844517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" event={"ID":"14f4be19-469c-4ae1-af53-da001a4d6d33","Type":"ContainerStarted","Data":"1fc8d3de5b4094c39efd379765a040ddf46ff276463e921cad0d3757c75b7f2c"} Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.844640 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.846981 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" event={"ID":"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae","Type":"ContainerStarted","Data":"e6f29452e8078d68876dd8db1b6d93f713fc9784fd30c674552a317e98b99d38"} Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.847021 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" event={"ID":"25bd5a4f-3d79-4b0c-817a-2f6770b8c8ae","Type":"ContainerStarted","Data":"3991a96241b3f22df9fe9e981bdae3bdd8d86b9a55dff691136c113e3198ac02"} Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.906809 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" podStartSLOduration=2.906793315 podStartE2EDuration="2.906793315s" podCreationTimestamp="2025-11-27 17:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:14:24.88214112 +0000 UTC m=+300.154598472" watchObservedRunningTime="2025-11-27 17:14:24.906793315 +0000 UTC m=+300.179250667" Nov 27 17:14:24 crc kubenswrapper[4809]: I1127 17:14:24.908613 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" podStartSLOduration=2.908602965 podStartE2EDuration="2.908602965s" podCreationTimestamp="2025-11-27 17:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:14:24.906756034 +0000 UTC m=+300.179213386" watchObservedRunningTime="2025-11-27 17:14:24.908602965 +0000 UTC m=+300.181060317" Nov 27 17:14:25 crc kubenswrapper[4809]: I1127 17:14:25.103080 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:25 crc kubenswrapper[4809]: I1127 17:14:25.237283 4809 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 27 17:14:25 crc kubenswrapper[4809]: I1127 17:14:25.851902 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:25 crc kubenswrapper[4809]: I1127 17:14:25.856561 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7bf57cbd5-ksszv" Nov 27 17:14:31 crc kubenswrapper[4809]: I1127 17:14:31.150381 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:14:32 crc kubenswrapper[4809]: I1127 17:14:32.127150 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:14:32 crc kubenswrapper[4809]: I1127 17:14:32.132227 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:14:39 crc kubenswrapper[4809]: I1127 17:14:39.797113 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68665d969-sctbd"] Nov 27 17:14:39 crc kubenswrapper[4809]: I1127 17:14:39.798950 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" podUID="14f4be19-469c-4ae1-af53-da001a4d6d33" containerName="route-controller-manager" containerID="cri-o://04f26916c0840859c8c2594175b57ab2ea0a80ea4325f959f8ef8a570a90c1eb" gracePeriod=30 Nov 27 17:14:39 crc kubenswrapper[4809]: I1127 17:14:39.947454 4809 generic.go:334] "Generic (PLEG): container finished" podID="14f4be19-469c-4ae1-af53-da001a4d6d33" containerID="04f26916c0840859c8c2594175b57ab2ea0a80ea4325f959f8ef8a570a90c1eb" exitCode=0 Nov 27 17:14:39 crc kubenswrapper[4809]: I1127 17:14:39.947656 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" event={"ID":"14f4be19-469c-4ae1-af53-da001a4d6d33","Type":"ContainerDied","Data":"04f26916c0840859c8c2594175b57ab2ea0a80ea4325f959f8ef8a570a90c1eb"} Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.282396 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.418128 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27rw9\" (UniqueName: \"kubernetes.io/projected/14f4be19-469c-4ae1-af53-da001a4d6d33-kube-api-access-27rw9\") pod \"14f4be19-469c-4ae1-af53-da001a4d6d33\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.418264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14f4be19-469c-4ae1-af53-da001a4d6d33-serving-cert\") pod \"14f4be19-469c-4ae1-af53-da001a4d6d33\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.418302 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-config\") pod \"14f4be19-469c-4ae1-af53-da001a4d6d33\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.418324 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-client-ca\") pod \"14f4be19-469c-4ae1-af53-da001a4d6d33\" (UID: \"14f4be19-469c-4ae1-af53-da001a4d6d33\") " Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.419240 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-client-ca" (OuterVolumeSpecName: "client-ca") pod "14f4be19-469c-4ae1-af53-da001a4d6d33" (UID: "14f4be19-469c-4ae1-af53-da001a4d6d33"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.421476 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-config" (OuterVolumeSpecName: "config") pod "14f4be19-469c-4ae1-af53-da001a4d6d33" (UID: "14f4be19-469c-4ae1-af53-da001a4d6d33"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.426548 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f4be19-469c-4ae1-af53-da001a4d6d33-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "14f4be19-469c-4ae1-af53-da001a4d6d33" (UID: "14f4be19-469c-4ae1-af53-da001a4d6d33"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.426970 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f4be19-469c-4ae1-af53-da001a4d6d33-kube-api-access-27rw9" (OuterVolumeSpecName: "kube-api-access-27rw9") pod "14f4be19-469c-4ae1-af53-da001a4d6d33" (UID: "14f4be19-469c-4ae1-af53-da001a4d6d33"). InnerVolumeSpecName "kube-api-access-27rw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.520523 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14f4be19-469c-4ae1-af53-da001a4d6d33-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.520570 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.520583 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14f4be19-469c-4ae1-af53-da001a4d6d33-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.520595 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27rw9\" (UniqueName: \"kubernetes.io/projected/14f4be19-469c-4ae1-af53-da001a4d6d33-kube-api-access-27rw9\") on node \"crc\" DevicePath \"\"" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.919799 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45"] Nov 27 17:14:40 crc kubenswrapper[4809]: E1127 17:14:40.920851 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f4be19-469c-4ae1-af53-da001a4d6d33" containerName="route-controller-manager" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.920873 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f4be19-469c-4ae1-af53-da001a4d6d33" containerName="route-controller-manager" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.921005 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f4be19-469c-4ae1-af53-da001a4d6d33" containerName="route-controller-manager" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.921676 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.932115 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45"] Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.957438 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" event={"ID":"14f4be19-469c-4ae1-af53-da001a4d6d33","Type":"ContainerDied","Data":"1fc8d3de5b4094c39efd379765a040ddf46ff276463e921cad0d3757c75b7f2c"} Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.957492 4809 scope.go:117] "RemoveContainer" containerID="04f26916c0840859c8c2594175b57ab2ea0a80ea4325f959f8ef8a570a90c1eb" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.957616 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68665d969-sctbd" Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.989876 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68665d969-sctbd"] Nov 27 17:14:40 crc kubenswrapper[4809]: I1127 17:14:40.996845 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68665d969-sctbd"] Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.027093 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-client-ca\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.027145 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-serving-cert\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.027182 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzhcd\" (UniqueName: \"kubernetes.io/projected/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-kube-api-access-qzhcd\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.027276 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-config\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.128915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-config\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.128973 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-client-ca\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.129000 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-serving-cert\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.129038 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzhcd\" (UniqueName: \"kubernetes.io/projected/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-kube-api-access-qzhcd\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.130295 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-client-ca\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.130348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-config\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.133934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-serving-cert\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.156030 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.168103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzhcd\" (UniqueName: \"kubernetes.io/projected/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-kube-api-access-qzhcd\") pod \"route-controller-manager-d49bbb4c4-ghc45\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.242450 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.465510 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14f4be19-469c-4ae1-af53-da001a4d6d33" path="/var/lib/kubelet/pods/14f4be19-469c-4ae1-af53-da001a4d6d33/volumes" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.663674 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45"] Nov 27 17:14:41 crc kubenswrapper[4809]: W1127 17:14:41.670376 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c7107e8_c8e2_4f0b_aa52_c447b342f3ea.slice/crio-cf1291a0743b8f4ec38e45074457467ccee4461008c717d653da4d1ed0b4bb85 WatchSource:0}: Error finding container cf1291a0743b8f4ec38e45074457467ccee4461008c717d653da4d1ed0b4bb85: Status 404 returned error can't find the container with id cf1291a0743b8f4ec38e45074457467ccee4461008c717d653da4d1ed0b4bb85 Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.964561 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" event={"ID":"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea","Type":"ContainerStarted","Data":"eabcb084f5cd0159cc93f72c5eadaf0a49a2ec3bba629ba13e5d9e8f19d17f14"} Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.964660 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" event={"ID":"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea","Type":"ContainerStarted","Data":"cf1291a0743b8f4ec38e45074457467ccee4461008c717d653da4d1ed0b4bb85"} Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.965813 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:14:41 crc kubenswrapper[4809]: I1127 17:14:41.989695 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" podStartSLOduration=2.989653793 podStartE2EDuration="2.989653793s" podCreationTimestamp="2025-11-27 17:14:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:14:41.98389644 +0000 UTC m=+317.256353832" watchObservedRunningTime="2025-11-27 17:14:41.989653793 +0000 UTC m=+317.262111145" Nov 27 17:14:42 crc kubenswrapper[4809]: I1127 17:14:42.225528 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.170706 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j"] Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.172905 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.176269 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.177652 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.180104 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j"] Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.193367 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdbfp\" (UniqueName: \"kubernetes.io/projected/b3f4af94-6c9c-498a-b1c9-828c7bce2226-kube-api-access-gdbfp\") pod \"collect-profiles-29404395-d698j\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.193492 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f4af94-6c9c-498a-b1c9-828c7bce2226-secret-volume\") pod \"collect-profiles-29404395-d698j\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.193577 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f4af94-6c9c-498a-b1c9-828c7bce2226-config-volume\") pod \"collect-profiles-29404395-d698j\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.294730 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f4af94-6c9c-498a-b1c9-828c7bce2226-secret-volume\") pod \"collect-profiles-29404395-d698j\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.294812 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f4af94-6c9c-498a-b1c9-828c7bce2226-config-volume\") pod \"collect-profiles-29404395-d698j\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.295142 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdbfp\" (UniqueName: \"kubernetes.io/projected/b3f4af94-6c9c-498a-b1c9-828c7bce2226-kube-api-access-gdbfp\") pod \"collect-profiles-29404395-d698j\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.296308 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f4af94-6c9c-498a-b1c9-828c7bce2226-config-volume\") pod \"collect-profiles-29404395-d698j\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.304690 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f4af94-6c9c-498a-b1c9-828c7bce2226-secret-volume\") pod \"collect-profiles-29404395-d698j\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.315410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdbfp\" (UniqueName: \"kubernetes.io/projected/b3f4af94-6c9c-498a-b1c9-828c7bce2226-kube-api-access-gdbfp\") pod \"collect-profiles-29404395-d698j\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.498396 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:00 crc kubenswrapper[4809]: I1127 17:15:00.908575 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j"] Nov 27 17:15:00 crc kubenswrapper[4809]: W1127 17:15:00.916480 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3f4af94_6c9c_498a_b1c9_828c7bce2226.slice/crio-b7b55f9ed69cca0635aad06ccc6749aaef6da589cae83d1586375a08ad587fbd WatchSource:0}: Error finding container b7b55f9ed69cca0635aad06ccc6749aaef6da589cae83d1586375a08ad587fbd: Status 404 returned error can't find the container with id b7b55f9ed69cca0635aad06ccc6749aaef6da589cae83d1586375a08ad587fbd Nov 27 17:15:01 crc kubenswrapper[4809]: I1127 17:15:01.091190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" event={"ID":"b3f4af94-6c9c-498a-b1c9-828c7bce2226","Type":"ContainerStarted","Data":"05a94a0b1200c4080367a07691589fd5a78843b330a83f0c054465ad4ec2f5c1"} Nov 27 17:15:01 crc kubenswrapper[4809]: I1127 17:15:01.091608 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" event={"ID":"b3f4af94-6c9c-498a-b1c9-828c7bce2226","Type":"ContainerStarted","Data":"b7b55f9ed69cca0635aad06ccc6749aaef6da589cae83d1586375a08ad587fbd"} Nov 27 17:15:01 crc kubenswrapper[4809]: I1127 17:15:01.115045 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" podStartSLOduration=1.115027799 podStartE2EDuration="1.115027799s" podCreationTimestamp="2025-11-27 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:15:01.11362816 +0000 UTC m=+336.386085502" watchObservedRunningTime="2025-11-27 17:15:01.115027799 +0000 UTC m=+336.387485151" Nov 27 17:15:02 crc kubenswrapper[4809]: I1127 17:15:02.097792 4809 generic.go:334] "Generic (PLEG): container finished" podID="b3f4af94-6c9c-498a-b1c9-828c7bce2226" containerID="05a94a0b1200c4080367a07691589fd5a78843b330a83f0c054465ad4ec2f5c1" exitCode=0 Nov 27 17:15:02 crc kubenswrapper[4809]: I1127 17:15:02.097832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" event={"ID":"b3f4af94-6c9c-498a-b1c9-828c7bce2226","Type":"ContainerDied","Data":"05a94a0b1200c4080367a07691589fd5a78843b330a83f0c054465ad4ec2f5c1"} Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.392626 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.535162 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f4af94-6c9c-498a-b1c9-828c7bce2226-secret-volume\") pod \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.535263 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdbfp\" (UniqueName: \"kubernetes.io/projected/b3f4af94-6c9c-498a-b1c9-828c7bce2226-kube-api-access-gdbfp\") pod \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.535357 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f4af94-6c9c-498a-b1c9-828c7bce2226-config-volume\") pod \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\" (UID: \"b3f4af94-6c9c-498a-b1c9-828c7bce2226\") " Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.536024 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f4af94-6c9c-498a-b1c9-828c7bce2226-config-volume" (OuterVolumeSpecName: "config-volume") pod "b3f4af94-6c9c-498a-b1c9-828c7bce2226" (UID: "b3f4af94-6c9c-498a-b1c9-828c7bce2226"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.540301 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3f4af94-6c9c-498a-b1c9-828c7bce2226-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b3f4af94-6c9c-498a-b1c9-828c7bce2226" (UID: "b3f4af94-6c9c-498a-b1c9-828c7bce2226"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.540828 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3f4af94-6c9c-498a-b1c9-828c7bce2226-kube-api-access-gdbfp" (OuterVolumeSpecName: "kube-api-access-gdbfp") pod "b3f4af94-6c9c-498a-b1c9-828c7bce2226" (UID: "b3f4af94-6c9c-498a-b1c9-828c7bce2226"). InnerVolumeSpecName "kube-api-access-gdbfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.636344 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdbfp\" (UniqueName: \"kubernetes.io/projected/b3f4af94-6c9c-498a-b1c9-828c7bce2226-kube-api-access-gdbfp\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.636384 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f4af94-6c9c-498a-b1c9-828c7bce2226-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:03 crc kubenswrapper[4809]: I1127 17:15:03.636416 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f4af94-6c9c-498a-b1c9-828c7bce2226-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:04 crc kubenswrapper[4809]: I1127 17:15:04.113312 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" event={"ID":"b3f4af94-6c9c-498a-b1c9-828c7bce2226","Type":"ContainerDied","Data":"b7b55f9ed69cca0635aad06ccc6749aaef6da589cae83d1586375a08ad587fbd"} Nov 27 17:15:04 crc kubenswrapper[4809]: I1127 17:15:04.113398 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7b55f9ed69cca0635aad06ccc6749aaef6da589cae83d1586375a08ad587fbd" Nov 27 17:15:04 crc kubenswrapper[4809]: I1127 17:15:04.113402 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.121638 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4chkp"] Nov 27 17:15:14 crc kubenswrapper[4809]: E1127 17:15:14.122843 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3f4af94-6c9c-498a-b1c9-828c7bce2226" containerName="collect-profiles" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.122862 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3f4af94-6c9c-498a-b1c9-828c7bce2226" containerName="collect-profiles" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.122984 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3f4af94-6c9c-498a-b1c9-828c7bce2226" containerName="collect-profiles" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.123501 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.146506 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4chkp"] Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.300151 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.300278 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-bound-sa-token\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.300308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-registry-tls\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.300331 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.300355 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94g9f\" (UniqueName: \"kubernetes.io/projected/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-kube-api-access-94g9f\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.300380 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-registry-certificates\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.300398 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.300652 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-trusted-ca\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.329077 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.402562 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-trusted-ca\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.402672 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-bound-sa-token\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.402698 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-registry-tls\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.402725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.402779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94g9f\" (UniqueName: \"kubernetes.io/projected/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-kube-api-access-94g9f\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.402806 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-registry-certificates\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.402829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.403589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.404224 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-registry-certificates\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.404268 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-trusted-ca\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.415901 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-registry-tls\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.416978 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.421729 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94g9f\" (UniqueName: \"kubernetes.io/projected/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-kube-api-access-94g9f\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.424193 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed3baf9d-bd92-4683-ac29-942f3e32b7e7-bound-sa-token\") pod \"image-registry-66df7c8f76-4chkp\" (UID: \"ed3baf9d-bd92-4683-ac29-942f3e32b7e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.445410 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:14 crc kubenswrapper[4809]: I1127 17:15:14.865136 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4chkp"] Nov 27 17:15:15 crc kubenswrapper[4809]: I1127 17:15:15.184008 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" event={"ID":"ed3baf9d-bd92-4683-ac29-942f3e32b7e7","Type":"ContainerStarted","Data":"e4cb36d8f67810844852562abc1260cf814a42ddcc4f70cf115890cb9a2944a6"} Nov 27 17:15:15 crc kubenswrapper[4809]: I1127 17:15:15.184062 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" event={"ID":"ed3baf9d-bd92-4683-ac29-942f3e32b7e7","Type":"ContainerStarted","Data":"d71268ab6248fc4845a33726e4a7436d82b70c70564a8eb0b2282b0979b5d894"} Nov 27 17:15:15 crc kubenswrapper[4809]: I1127 17:15:15.185015 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:15 crc kubenswrapper[4809]: I1127 17:15:15.216235 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" podStartSLOduration=1.216195425 podStartE2EDuration="1.216195425s" podCreationTimestamp="2025-11-27 17:15:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:15:15.212335929 +0000 UTC m=+350.484793291" watchObservedRunningTime="2025-11-27 17:15:15.216195425 +0000 UTC m=+350.488652797" Nov 27 17:15:21 crc kubenswrapper[4809]: I1127 17:15:21.974599 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45"] Nov 27 17:15:21 crc kubenswrapper[4809]: I1127 17:15:21.975960 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" podUID="5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" containerName="route-controller-manager" containerID="cri-o://eabcb084f5cd0159cc93f72c5eadaf0a49a2ec3bba629ba13e5d9e8f19d17f14" gracePeriod=30 Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.230915 4809 generic.go:334] "Generic (PLEG): container finished" podID="5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" containerID="eabcb084f5cd0159cc93f72c5eadaf0a49a2ec3bba629ba13e5d9e8f19d17f14" exitCode=0 Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.230986 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" event={"ID":"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea","Type":"ContainerDied","Data":"eabcb084f5cd0159cc93f72c5eadaf0a49a2ec3bba629ba13e5d9e8f19d17f14"} Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.566014 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.745686 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-serving-cert\") pod \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.745905 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-client-ca\") pod \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.745975 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzhcd\" (UniqueName: \"kubernetes.io/projected/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-kube-api-access-qzhcd\") pod \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.746106 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-config\") pod \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\" (UID: \"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea\") " Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.747036 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-config" (OuterVolumeSpecName: "config") pod "5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" (UID: "5c7107e8-c8e2-4f0b-aa52-c447b342f3ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.747376 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-client-ca" (OuterVolumeSpecName: "client-ca") pod "5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" (UID: "5c7107e8-c8e2-4f0b-aa52-c447b342f3ea"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.752846 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" (UID: "5c7107e8-c8e2-4f0b-aa52-c447b342f3ea"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.753448 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-kube-api-access-qzhcd" (OuterVolumeSpecName: "kube-api-access-qzhcd") pod "5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" (UID: "5c7107e8-c8e2-4f0b-aa52-c447b342f3ea"). InnerVolumeSpecName "kube-api-access-qzhcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.847466 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.847511 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.847522 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzhcd\" (UniqueName: \"kubernetes.io/projected/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-kube-api-access-qzhcd\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:22 crc kubenswrapper[4809]: I1127 17:15:22.847531 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.239225 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" event={"ID":"5c7107e8-c8e2-4f0b-aa52-c447b342f3ea","Type":"ContainerDied","Data":"cf1291a0743b8f4ec38e45074457467ccee4461008c717d653da4d1ed0b4bb85"} Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.239303 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.239890 4809 scope.go:117] "RemoveContainer" containerID="eabcb084f5cd0159cc93f72c5eadaf0a49a2ec3bba629ba13e5d9e8f19d17f14" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.269108 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45"] Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.273599 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d49bbb4c4-ghc45"] Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.464504 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" path="/var/lib/kubelet/pods/5c7107e8-c8e2-4f0b-aa52-c447b342f3ea/volumes" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.951432 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68665d969-t4f25"] Nov 27 17:15:23 crc kubenswrapper[4809]: E1127 17:15:23.951763 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" containerName="route-controller-manager" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.951778 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" containerName="route-controller-manager" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.951886 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c7107e8-c8e2-4f0b-aa52-c447b342f3ea" containerName="route-controller-manager" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.952353 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.954273 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.954938 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.955800 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.956326 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.956363 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.956603 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 17:15:23 crc kubenswrapper[4809]: I1127 17:15:23.962718 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68665d969-t4f25"] Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.065518 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a76c4222-c6fc-4d20-8223-a026bbe0acee-config\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.065561 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a76c4222-c6fc-4d20-8223-a026bbe0acee-serving-cert\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.065594 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a76c4222-c6fc-4d20-8223-a026bbe0acee-client-ca\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.065663 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l79jf\" (UniqueName: \"kubernetes.io/projected/a76c4222-c6fc-4d20-8223-a026bbe0acee-kube-api-access-l79jf\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.166873 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a76c4222-c6fc-4d20-8223-a026bbe0acee-config\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.166939 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a76c4222-c6fc-4d20-8223-a026bbe0acee-serving-cert\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.166970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a76c4222-c6fc-4d20-8223-a026bbe0acee-client-ca\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.167033 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l79jf\" (UniqueName: \"kubernetes.io/projected/a76c4222-c6fc-4d20-8223-a026bbe0acee-kube-api-access-l79jf\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.168664 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a76c4222-c6fc-4d20-8223-a026bbe0acee-client-ca\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.169111 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a76c4222-c6fc-4d20-8223-a026bbe0acee-config\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.173072 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a76c4222-c6fc-4d20-8223-a026bbe0acee-serving-cert\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.184216 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l79jf\" (UniqueName: \"kubernetes.io/projected/a76c4222-c6fc-4d20-8223-a026bbe0acee-kube-api-access-l79jf\") pod \"route-controller-manager-68665d969-t4f25\" (UID: \"a76c4222-c6fc-4d20-8223-a026bbe0acee\") " pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.277723 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:24 crc kubenswrapper[4809]: I1127 17:15:24.485741 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68665d969-t4f25"] Nov 27 17:15:25 crc kubenswrapper[4809]: I1127 17:15:25.258714 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" event={"ID":"a76c4222-c6fc-4d20-8223-a026bbe0acee","Type":"ContainerStarted","Data":"102e93e6a55fbfbd6c05f4e244768510b03af7f1cc4714d9711765b8a6f6cd06"} Nov 27 17:15:25 crc kubenswrapper[4809]: I1127 17:15:25.259108 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:25 crc kubenswrapper[4809]: I1127 17:15:25.259138 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" event={"ID":"a76c4222-c6fc-4d20-8223-a026bbe0acee","Type":"ContainerStarted","Data":"f4ce02ff175a30da17a0a8668fe04e210d41afd8ec61fcf0d6c8b911ab729922"} Nov 27 17:15:25 crc kubenswrapper[4809]: I1127 17:15:25.271216 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" Nov 27 17:15:25 crc kubenswrapper[4809]: I1127 17:15:25.284455 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-68665d969-t4f25" podStartSLOduration=4.28443553 podStartE2EDuration="4.28443553s" podCreationTimestamp="2025-11-27 17:15:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:15:25.283090303 +0000 UTC m=+360.555547655" watchObservedRunningTime="2025-11-27 17:15:25.28443553 +0000 UTC m=+360.556892892" Nov 27 17:15:25 crc kubenswrapper[4809]: I1127 17:15:25.780310 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:15:25 crc kubenswrapper[4809]: I1127 17:15:25.781044 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.185955 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7hk4l"] Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.187335 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7hk4l" podUID="03562c92-cba5-4c9a-b759-7da40962be62" containerName="registry-server" containerID="cri-o://a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493" gracePeriod=30 Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.193530 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrw24"] Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.193777 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hrw24" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" containerName="registry-server" containerID="cri-o://923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd" gracePeriod=30 Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.208136 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-94nbr"] Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.208371 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" podUID="92093439-c71b-4fdd-88ff-189ecd253269" containerName="marketplace-operator" containerID="cri-o://25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb" gracePeriod=30 Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.216123 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-chw9h"] Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.216435 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-chw9h" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" containerName="registry-server" containerID="cri-o://1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28" gracePeriod=30 Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.230108 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-64fv9"] Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.232073 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.242471 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dcvzh"] Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.242841 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dcvzh" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerName="registry-server" containerID="cri-o://92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a" gracePeriod=30 Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.243936 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-64fv9"] Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.358102 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8szx\" (UniqueName: \"kubernetes.io/projected/f2bae280-204f-4f0f-93f5-fd574d9d4b79-kube-api-access-z8szx\") pod \"marketplace-operator-79b997595-64fv9\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.358383 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-64fv9\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.358409 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-64fv9\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: E1127 17:15:30.400168 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28 is running failed: container process not found" containerID="1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 17:15:30 crc kubenswrapper[4809]: E1127 17:15:30.400581 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28 is running failed: container process not found" containerID="1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 17:15:30 crc kubenswrapper[4809]: E1127 17:15:30.400843 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28 is running failed: container process not found" containerID="1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 17:15:30 crc kubenswrapper[4809]: E1127 17:15:30.400884 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-chw9h" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" containerName="registry-server" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.459345 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8szx\" (UniqueName: \"kubernetes.io/projected/f2bae280-204f-4f0f-93f5-fd574d9d4b79-kube-api-access-z8szx\") pod \"marketplace-operator-79b997595-64fv9\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.459407 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-64fv9\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.459436 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-64fv9\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.460596 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-64fv9\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.470630 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-64fv9\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.477610 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8szx\" (UniqueName: \"kubernetes.io/projected/f2bae280-204f-4f0f-93f5-fd574d9d4b79-kube-api-access-z8szx\") pod \"marketplace-operator-79b997595-64fv9\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.518443 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.561142 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.660818 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-utilities\") pod \"b1d9678d-171a-427b-9b37-192f523e63b2\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.661414 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8g96\" (UniqueName: \"kubernetes.io/projected/b1d9678d-171a-427b-9b37-192f523e63b2-kube-api-access-x8g96\") pod \"b1d9678d-171a-427b-9b37-192f523e63b2\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.661462 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-catalog-content\") pod \"b1d9678d-171a-427b-9b37-192f523e63b2\" (UID: \"b1d9678d-171a-427b-9b37-192f523e63b2\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.662198 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-utilities" (OuterVolumeSpecName: "utilities") pod "b1d9678d-171a-427b-9b37-192f523e63b2" (UID: "b1d9678d-171a-427b-9b37-192f523e63b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.667601 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1d9678d-171a-427b-9b37-192f523e63b2-kube-api-access-x8g96" (OuterVolumeSpecName: "kube-api-access-x8g96") pod "b1d9678d-171a-427b-9b37-192f523e63b2" (UID: "b1d9678d-171a-427b-9b37-192f523e63b2"). InnerVolumeSpecName "kube-api-access-x8g96". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.668894 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.672949 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.737572 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.738565 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1d9678d-171a-427b-9b37-192f523e63b2" (UID: "b1d9678d-171a-427b-9b37-192f523e63b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.748197 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.766587 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtnx6\" (UniqueName: \"kubernetes.io/projected/867f6bd0-b172-4409-a736-cfa7484d5685-kube-api-access-gtnx6\") pod \"867f6bd0-b172-4409-a736-cfa7484d5685\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.766650 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-utilities\") pod \"867f6bd0-b172-4409-a736-cfa7484d5685\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.766714 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-catalog-content\") pod \"867f6bd0-b172-4409-a736-cfa7484d5685\" (UID: \"867f6bd0-b172-4409-a736-cfa7484d5685\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.767015 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8g96\" (UniqueName: \"kubernetes.io/projected/b1d9678d-171a-427b-9b37-192f523e63b2-kube-api-access-x8g96\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.767038 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.767048 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9678d-171a-427b-9b37-192f523e63b2-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.772959 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/867f6bd0-b172-4409-a736-cfa7484d5685-kube-api-access-gtnx6" (OuterVolumeSpecName: "kube-api-access-gtnx6") pod "867f6bd0-b172-4409-a736-cfa7484d5685" (UID: "867f6bd0-b172-4409-a736-cfa7484d5685"). InnerVolumeSpecName "kube-api-access-gtnx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.778667 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-utilities" (OuterVolumeSpecName: "utilities") pod "867f6bd0-b172-4409-a736-cfa7484d5685" (UID: "867f6bd0-b172-4409-a736-cfa7484d5685"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.787551 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "867f6bd0-b172-4409-a736-cfa7484d5685" (UID: "867f6bd0-b172-4409-a736-cfa7484d5685"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.868707 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhbmt\" (UniqueName: \"kubernetes.io/projected/92093439-c71b-4fdd-88ff-189ecd253269-kube-api-access-zhbmt\") pod \"92093439-c71b-4fdd-88ff-189ecd253269\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.868796 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-trusted-ca\") pod \"92093439-c71b-4fdd-88ff-189ecd253269\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.868820 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2776\" (UniqueName: \"kubernetes.io/projected/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-kube-api-access-l2776\") pod \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.868872 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-utilities\") pod \"03562c92-cba5-4c9a-b759-7da40962be62\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.868952 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-catalog-content\") pod \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.868979 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-utilities\") pod \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\" (UID: \"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869002 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-catalog-content\") pod \"03562c92-cba5-4c9a-b759-7da40962be62\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869059 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-operator-metrics\") pod \"92093439-c71b-4fdd-88ff-189ecd253269\" (UID: \"92093439-c71b-4fdd-88ff-189ecd253269\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869079 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2qzw\" (UniqueName: \"kubernetes.io/projected/03562c92-cba5-4c9a-b759-7da40962be62-kube-api-access-v2qzw\") pod \"03562c92-cba5-4c9a-b759-7da40962be62\" (UID: \"03562c92-cba5-4c9a-b759-7da40962be62\") " Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869253 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "92093439-c71b-4fdd-88ff-189ecd253269" (UID: "92093439-c71b-4fdd-88ff-189ecd253269"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869394 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtnx6\" (UniqueName: \"kubernetes.io/projected/867f6bd0-b172-4409-a736-cfa7484d5685-kube-api-access-gtnx6\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869431 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869441 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869451 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/867f6bd0-b172-4409-a736-cfa7484d5685-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869560 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-utilities" (OuterVolumeSpecName: "utilities") pod "03562c92-cba5-4c9a-b759-7da40962be62" (UID: "03562c92-cba5-4c9a-b759-7da40962be62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.869902 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-utilities" (OuterVolumeSpecName: "utilities") pod "225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" (UID: "225a59f5-d662-4f63-8ec6-c7f3eeb6ba40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.871705 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92093439-c71b-4fdd-88ff-189ecd253269-kube-api-access-zhbmt" (OuterVolumeSpecName: "kube-api-access-zhbmt") pod "92093439-c71b-4fdd-88ff-189ecd253269" (UID: "92093439-c71b-4fdd-88ff-189ecd253269"). InnerVolumeSpecName "kube-api-access-zhbmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.872430 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "92093439-c71b-4fdd-88ff-189ecd253269" (UID: "92093439-c71b-4fdd-88ff-189ecd253269"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.872500 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-kube-api-access-l2776" (OuterVolumeSpecName: "kube-api-access-l2776") pod "225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" (UID: "225a59f5-d662-4f63-8ec6-c7f3eeb6ba40"). InnerVolumeSpecName "kube-api-access-l2776". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.872702 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03562c92-cba5-4c9a-b759-7da40962be62-kube-api-access-v2qzw" (OuterVolumeSpecName: "kube-api-access-v2qzw") pod "03562c92-cba5-4c9a-b759-7da40962be62" (UID: "03562c92-cba5-4c9a-b759-7da40962be62"). InnerVolumeSpecName "kube-api-access-v2qzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.917062 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03562c92-cba5-4c9a-b759-7da40962be62" (UID: "03562c92-cba5-4c9a-b759-7da40962be62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:15:30 crc kubenswrapper[4809]: W1127 17:15:30.960154 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2bae280_204f_4f0f_93f5_fd574d9d4b79.slice/crio-33c084717ecee0daeb906fd16edeac01ba603b9a8692bd529b2e29abc8506285 WatchSource:0}: Error finding container 33c084717ecee0daeb906fd16edeac01ba603b9a8692bd529b2e29abc8506285: Status 404 returned error can't find the container with id 33c084717ecee0daeb906fd16edeac01ba603b9a8692bd529b2e29abc8506285 Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.962798 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-64fv9"] Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.971017 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.971045 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.971056 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/92093439-c71b-4fdd-88ff-189ecd253269-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.971066 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2qzw\" (UniqueName: \"kubernetes.io/projected/03562c92-cba5-4c9a-b759-7da40962be62-kube-api-access-v2qzw\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.971075 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhbmt\" (UniqueName: \"kubernetes.io/projected/92093439-c71b-4fdd-88ff-189ecd253269-kube-api-access-zhbmt\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.971083 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2776\" (UniqueName: \"kubernetes.io/projected/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-kube-api-access-l2776\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.971091 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03562c92-cba5-4c9a-b759-7da40962be62-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:30 crc kubenswrapper[4809]: I1127 17:15:30.982873 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" (UID: "225a59f5-d662-4f63-8ec6-c7f3eeb6ba40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.072102 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.291988 4809 generic.go:334] "Generic (PLEG): container finished" podID="867f6bd0-b172-4409-a736-cfa7484d5685" containerID="1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28" exitCode=0 Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.292054 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-chw9h" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.292058 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chw9h" event={"ID":"867f6bd0-b172-4409-a736-cfa7484d5685","Type":"ContainerDied","Data":"1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.292086 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-chw9h" event={"ID":"867f6bd0-b172-4409-a736-cfa7484d5685","Type":"ContainerDied","Data":"50b824330a9c147fc0aa98c1ded4ef065546b1504561abcf6c9b11beb76e0685"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.292114 4809 scope.go:117] "RemoveContainer" containerID="1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.295331 4809 generic.go:334] "Generic (PLEG): container finished" podID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerID="92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a" exitCode=0 Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.295408 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dcvzh" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.295399 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcvzh" event={"ID":"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40","Type":"ContainerDied","Data":"92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.295915 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dcvzh" event={"ID":"225a59f5-d662-4f63-8ec6-c7f3eeb6ba40","Type":"ContainerDied","Data":"f383b83a837a0b9f10fef1809bd65baccb1f83e237d9bc21db8c0b06cd6df181"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.296765 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" event={"ID":"92093439-c71b-4fdd-88ff-189ecd253269","Type":"ContainerDied","Data":"25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.296727 4809 generic.go:334] "Generic (PLEG): container finished" podID="92093439-c71b-4fdd-88ff-189ecd253269" containerID="25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb" exitCode=0 Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.296843 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.296848 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-94nbr" event={"ID":"92093439-c71b-4fdd-88ff-189ecd253269","Type":"ContainerDied","Data":"9690739b69eb1529ffbfe37ba028e444e46a4cbe247f1f23359ca9f7f8abd171"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.299991 4809 generic.go:334] "Generic (PLEG): container finished" podID="b1d9678d-171a-427b-9b37-192f523e63b2" containerID="923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd" exitCode=0 Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.300060 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrw24" event={"ID":"b1d9678d-171a-427b-9b37-192f523e63b2","Type":"ContainerDied","Data":"923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.300085 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrw24" event={"ID":"b1d9678d-171a-427b-9b37-192f523e63b2","Type":"ContainerDied","Data":"bf9d146f95ef73133aca3a81ccf5559f8f6a24391f18a43aa1d9c1388a1497b9"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.300045 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrw24" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.301926 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" event={"ID":"f2bae280-204f-4f0f-93f5-fd574d9d4b79","Type":"ContainerStarted","Data":"f3afb48367da35625b2b19752ebc9d08c76f8857f69cfa18e56a1c71f8ccf2d2"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.301962 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" event={"ID":"f2bae280-204f-4f0f-93f5-fd574d9d4b79","Type":"ContainerStarted","Data":"33c084717ecee0daeb906fd16edeac01ba603b9a8692bd529b2e29abc8506285"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.302368 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.303781 4809 generic.go:334] "Generic (PLEG): container finished" podID="03562c92-cba5-4c9a-b759-7da40962be62" containerID="a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493" exitCode=0 Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.303826 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hk4l" event={"ID":"03562c92-cba5-4c9a-b759-7da40962be62","Type":"ContainerDied","Data":"a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.303845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hk4l" event={"ID":"03562c92-cba5-4c9a-b759-7da40962be62","Type":"ContainerDied","Data":"16d65a77c932312d8c53ee42bccedbc803f51f53d2bc3dec491a5924d98a8f20"} Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.303894 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7hk4l" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.306145 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-64fv9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.64:8080/healthz\": dial tcp 10.217.0.64:8080: connect: connection refused" start-of-body= Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.306479 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" podUID="f2bae280-204f-4f0f-93f5-fd574d9d4b79" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.64:8080/healthz\": dial tcp 10.217.0.64:8080: connect: connection refused" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.307046 4809 scope.go:117] "RemoveContainer" containerID="1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.333832 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" podStartSLOduration=1.333810576 podStartE2EDuration="1.333810576s" podCreationTimestamp="2025-11-27 17:15:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:15:31.332108159 +0000 UTC m=+366.604565521" watchObservedRunningTime="2025-11-27 17:15:31.333810576 +0000 UTC m=+366.606267928" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.341270 4809 scope.go:117] "RemoveContainer" containerID="a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.354902 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-chw9h"] Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.363059 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-chw9h"] Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.368693 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-94nbr"] Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.374041 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-94nbr"] Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.379098 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dcvzh"] Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.388920 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dcvzh"] Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.403483 4809 scope.go:117] "RemoveContainer" containerID="1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.404122 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28\": container with ID starting with 1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28 not found: ID does not exist" containerID="1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.404241 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28"} err="failed to get container status \"1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28\": rpc error: code = NotFound desc = could not find container \"1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28\": container with ID starting with 1c2a3eb836191d48f007a983ab61cd24e74ec8422d832975b6a77ef1edf0ee28 not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.404374 4809 scope.go:117] "RemoveContainer" containerID="1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.405835 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66\": container with ID starting with 1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66 not found: ID does not exist" containerID="1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.405870 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66"} err="failed to get container status \"1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66\": rpc error: code = NotFound desc = could not find container \"1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66\": container with ID starting with 1d3e5b29290d1eee8b5ddf1eeede6f35dc9f2c3de9257f6fd11bb1b840872e66 not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.405891 4809 scope.go:117] "RemoveContainer" containerID="a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.406136 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrw24"] Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.406230 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a\": container with ID starting with a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a not found: ID does not exist" containerID="a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.406325 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a"} err="failed to get container status \"a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a\": rpc error: code = NotFound desc = could not find container \"a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a\": container with ID starting with a843cc256f1e41a02776de10683bdd9d8c9b5c3f894018bb9d087f9a25ddc83a not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.406424 4809 scope.go:117] "RemoveContainer" containerID="92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.412298 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hrw24"] Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.417030 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7hk4l"] Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.422780 4809 scope.go:117] "RemoveContainer" containerID="871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.425933 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7hk4l"] Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.438179 4809 scope.go:117] "RemoveContainer" containerID="34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.453305 4809 scope.go:117] "RemoveContainer" containerID="92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.453712 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a\": container with ID starting with 92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a not found: ID does not exist" containerID="92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.453762 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a"} err="failed to get container status \"92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a\": rpc error: code = NotFound desc = could not find container \"92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a\": container with ID starting with 92fffa4c9cd718dc9de6284c17eaf0ec676e60441ff72287004d8e7d4c6ce12a not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.453801 4809 scope.go:117] "RemoveContainer" containerID="871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.454393 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793\": container with ID starting with 871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793 not found: ID does not exist" containerID="871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.454413 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793"} err="failed to get container status \"871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793\": rpc error: code = NotFound desc = could not find container \"871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793\": container with ID starting with 871c05d9eb29d828c5ab25b6857c6412a685ea18fb77b2f956837de297d90793 not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.454427 4809 scope.go:117] "RemoveContainer" containerID="34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.454764 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6\": container with ID starting with 34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6 not found: ID does not exist" containerID="34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.454813 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6"} err="failed to get container status \"34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6\": rpc error: code = NotFound desc = could not find container \"34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6\": container with ID starting with 34a25269682efb94a5721dc8a6d74db1e85a54c79b95213bd2170d25bde36bb6 not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.454849 4809 scope.go:117] "RemoveContainer" containerID="25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.464893 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03562c92-cba5-4c9a-b759-7da40962be62" path="/var/lib/kubelet/pods/03562c92-cba5-4c9a-b759-7da40962be62/volumes" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.465583 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" path="/var/lib/kubelet/pods/225a59f5-d662-4f63-8ec6-c7f3eeb6ba40/volumes" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.466369 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" path="/var/lib/kubelet/pods/867f6bd0-b172-4409-a736-cfa7484d5685/volumes" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.467683 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92093439-c71b-4fdd-88ff-189ecd253269" path="/var/lib/kubelet/pods/92093439-c71b-4fdd-88ff-189ecd253269/volumes" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.468277 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" path="/var/lib/kubelet/pods/b1d9678d-171a-427b-9b37-192f523e63b2/volumes" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.471413 4809 scope.go:117] "RemoveContainer" containerID="25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.471895 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb\": container with ID starting with 25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb not found: ID does not exist" containerID="25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.471939 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb"} err="failed to get container status \"25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb\": rpc error: code = NotFound desc = could not find container \"25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb\": container with ID starting with 25dceaf0c2e8412a496861316783c52f15c7d4d0f940d2db6e204779ea310efb not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.471966 4809 scope.go:117] "RemoveContainer" containerID="923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.492511 4809 scope.go:117] "RemoveContainer" containerID="528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.510937 4809 scope.go:117] "RemoveContainer" containerID="14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.524247 4809 scope.go:117] "RemoveContainer" containerID="923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.524778 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd\": container with ID starting with 923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd not found: ID does not exist" containerID="923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.524814 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd"} err="failed to get container status \"923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd\": rpc error: code = NotFound desc = could not find container \"923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd\": container with ID starting with 923ee2c2ded6676369cc0bc6e178da3db81d8625f19e211367a92330a9ca48fd not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.524835 4809 scope.go:117] "RemoveContainer" containerID="528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.525044 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d\": container with ID starting with 528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d not found: ID does not exist" containerID="528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.525073 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d"} err="failed to get container status \"528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d\": rpc error: code = NotFound desc = could not find container \"528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d\": container with ID starting with 528d7911de3bfdd0d9314f23e6fd6f1fbcc926a15de294e05557015832f2532d not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.525090 4809 scope.go:117] "RemoveContainer" containerID="14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.525356 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca\": container with ID starting with 14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca not found: ID does not exist" containerID="14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.525382 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca"} err="failed to get container status \"14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca\": rpc error: code = NotFound desc = could not find container \"14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca\": container with ID starting with 14279eb845475cff92722f0339be309e24e5f3e59cc2f1a28f130292d1ca70ca not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.525400 4809 scope.go:117] "RemoveContainer" containerID="a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.536456 4809 scope.go:117] "RemoveContainer" containerID="7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.548707 4809 scope.go:117] "RemoveContainer" containerID="592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.562255 4809 scope.go:117] "RemoveContainer" containerID="a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.562705 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493\": container with ID starting with a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493 not found: ID does not exist" containerID="a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.562764 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493"} err="failed to get container status \"a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493\": rpc error: code = NotFound desc = could not find container \"a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493\": container with ID starting with a94bdbed394f2e2d94021385cdc7b6a56b68e3d64ca02d9b2d8107718d1d9493 not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.562796 4809 scope.go:117] "RemoveContainer" containerID="7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.563125 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f\": container with ID starting with 7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f not found: ID does not exist" containerID="7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.563149 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f"} err="failed to get container status \"7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f\": rpc error: code = NotFound desc = could not find container \"7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f\": container with ID starting with 7750b7b3186a8b256da15a69ba2368f3ccd96c3c9b29cdf9ccd39918c12c696f not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.563162 4809 scope.go:117] "RemoveContainer" containerID="592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.563588 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8\": container with ID starting with 592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8 not found: ID does not exist" containerID="592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.563617 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8"} err="failed to get container status \"592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8\": rpc error: code = NotFound desc = could not find container \"592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8\": container with ID starting with 592ac687900b747c93dd7da03cb03fa4b485774299a13f0058f7934843823ca8 not found: ID does not exist" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976141 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pgtgl"] Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976353 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" containerName="extract-utilities" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976365 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" containerName="extract-utilities" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976377 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92093439-c71b-4fdd-88ff-189ecd253269" containerName="marketplace-operator" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976383 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="92093439-c71b-4fdd-88ff-189ecd253269" containerName="marketplace-operator" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976393 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" containerName="extract-content" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976399 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" containerName="extract-content" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976407 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" containerName="extract-utilities" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976413 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" containerName="extract-utilities" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976421 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerName="extract-content" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976427 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerName="extract-content" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976435 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" containerName="extract-content" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976441 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" containerName="extract-content" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976452 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976457 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976466 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03562c92-cba5-4c9a-b759-7da40962be62" containerName="extract-utilities" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976472 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="03562c92-cba5-4c9a-b759-7da40962be62" containerName="extract-utilities" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976480 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03562c92-cba5-4c9a-b759-7da40962be62" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976487 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="03562c92-cba5-4c9a-b759-7da40962be62" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976496 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03562c92-cba5-4c9a-b759-7da40962be62" containerName="extract-content" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976502 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="03562c92-cba5-4c9a-b759-7da40962be62" containerName="extract-content" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976510 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976516 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976526 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976532 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: E1127 17:15:31.976539 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerName="extract-utilities" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976544 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerName="extract-utilities" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976620 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="225a59f5-d662-4f63-8ec6-c7f3eeb6ba40" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976629 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="92093439-c71b-4fdd-88ff-189ecd253269" containerName="marketplace-operator" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976639 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="867f6bd0-b172-4409-a736-cfa7484d5685" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976648 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="03562c92-cba5-4c9a-b759-7da40962be62" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.976656 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d9678d-171a-427b-9b37-192f523e63b2" containerName="registry-server" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.977351 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.979517 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.985951 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-utilities\") pod \"certified-operators-pgtgl\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.985995 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmnwt\" (UniqueName: \"kubernetes.io/projected/ceaa3143-e2dc-40fd-8100-253811eeb357-kube-api-access-hmnwt\") pod \"certified-operators-pgtgl\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.986019 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-catalog-content\") pod \"certified-operators-pgtgl\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:31 crc kubenswrapper[4809]: I1127 17:15:31.990230 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pgtgl"] Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.087971 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-utilities\") pod \"certified-operators-pgtgl\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.088025 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmnwt\" (UniqueName: \"kubernetes.io/projected/ceaa3143-e2dc-40fd-8100-253811eeb357-kube-api-access-hmnwt\") pod \"certified-operators-pgtgl\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.088048 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-catalog-content\") pod \"certified-operators-pgtgl\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.088436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-utilities\") pod \"certified-operators-pgtgl\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.088458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-catalog-content\") pod \"certified-operators-pgtgl\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.106236 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmnwt\" (UniqueName: \"kubernetes.io/projected/ceaa3143-e2dc-40fd-8100-253811eeb357-kube-api-access-hmnwt\") pod \"certified-operators-pgtgl\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.177271 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-swvpz"] Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.182968 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swvpz"] Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.183066 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.185187 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.188515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-utilities\") pod \"community-operators-swvpz\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.188548 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qsxn\" (UniqueName: \"kubernetes.io/projected/57211ee0-a59b-4c8b-99d0-6854e00b9e53-kube-api-access-7qsxn\") pod \"community-operators-swvpz\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.188589 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-catalog-content\") pod \"community-operators-swvpz\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.289904 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-catalog-content\") pod \"community-operators-swvpz\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.290004 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-utilities\") pod \"community-operators-swvpz\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.290039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qsxn\" (UniqueName: \"kubernetes.io/projected/57211ee0-a59b-4c8b-99d0-6854e00b9e53-kube-api-access-7qsxn\") pod \"community-operators-swvpz\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.290505 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-utilities\") pod \"community-operators-swvpz\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.290558 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-catalog-content\") pod \"community-operators-swvpz\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.297806 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.308031 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qsxn\" (UniqueName: \"kubernetes.io/projected/57211ee0-a59b-4c8b-99d0-6854e00b9e53-kube-api-access-7qsxn\") pod \"community-operators-swvpz\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.319048 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.499853 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.512345 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pgtgl"] Nov 27 17:15:32 crc kubenswrapper[4809]: I1127 17:15:32.890255 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swvpz"] Nov 27 17:15:32 crc kubenswrapper[4809]: W1127 17:15:32.899886 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57211ee0_a59b_4c8b_99d0_6854e00b9e53.slice/crio-73e2ca9389ce5c5c71530acb183a13806133870e028765585d894ac036217af3 WatchSource:0}: Error finding container 73e2ca9389ce5c5c71530acb183a13806133870e028765585d894ac036217af3: Status 404 returned error can't find the container with id 73e2ca9389ce5c5c71530acb183a13806133870e028765585d894ac036217af3 Nov 27 17:15:33 crc kubenswrapper[4809]: I1127 17:15:33.319959 4809 generic.go:334] "Generic (PLEG): container finished" podID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerID="27823b4e69930199d9cb28f54256cd3eef8f1f7ea84e8613fc1db3178a339ec1" exitCode=0 Nov 27 17:15:33 crc kubenswrapper[4809]: I1127 17:15:33.320026 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgtgl" event={"ID":"ceaa3143-e2dc-40fd-8100-253811eeb357","Type":"ContainerDied","Data":"27823b4e69930199d9cb28f54256cd3eef8f1f7ea84e8613fc1db3178a339ec1"} Nov 27 17:15:33 crc kubenswrapper[4809]: I1127 17:15:33.320337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgtgl" event={"ID":"ceaa3143-e2dc-40fd-8100-253811eeb357","Type":"ContainerStarted","Data":"b757815544d05d9b7f2872f8a330c1f600016a3da560fe9325558f8489584f72"} Nov 27 17:15:33 crc kubenswrapper[4809]: I1127 17:15:33.321573 4809 generic.go:334] "Generic (PLEG): container finished" podID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerID="42e863447013efda8e741ebac2812c74c25b3ba5dd50d509aad99f6d4443019e" exitCode=0 Nov 27 17:15:33 crc kubenswrapper[4809]: I1127 17:15:33.321690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvpz" event={"ID":"57211ee0-a59b-4c8b-99d0-6854e00b9e53","Type":"ContainerDied","Data":"42e863447013efda8e741ebac2812c74c25b3ba5dd50d509aad99f6d4443019e"} Nov 27 17:15:33 crc kubenswrapper[4809]: I1127 17:15:33.321717 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvpz" event={"ID":"57211ee0-a59b-4c8b-99d0-6854e00b9e53","Type":"ContainerStarted","Data":"73e2ca9389ce5c5c71530acb183a13806133870e028765585d894ac036217af3"} Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.328040 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgtgl" event={"ID":"ceaa3143-e2dc-40fd-8100-253811eeb357","Type":"ContainerStarted","Data":"fb3871a99fc82b18aca4345cb5e6d6a47649302dd57bf9409aa9dea99a8ee064"} Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.330184 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvpz" event={"ID":"57211ee0-a59b-4c8b-99d0-6854e00b9e53","Type":"ContainerStarted","Data":"5c459bb53c76e8f422b63e4b290c124a84aa77ee269dca530d24c6f01585f3e2"} Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.379132 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fvw4b"] Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.380430 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.383117 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.394432 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvw4b"] Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.415293 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-utilities\") pod \"redhat-marketplace-fvw4b\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.415344 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-catalog-content\") pod \"redhat-marketplace-fvw4b\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.415470 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtq8v\" (UniqueName: \"kubernetes.io/projected/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-kube-api-access-rtq8v\") pod \"redhat-marketplace-fvw4b\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.452797 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4chkp" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.508117 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dj24l"] Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.517696 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtq8v\" (UniqueName: \"kubernetes.io/projected/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-kube-api-access-rtq8v\") pod \"redhat-marketplace-fvw4b\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.517818 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-utilities\") pod \"redhat-marketplace-fvw4b\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.517848 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-catalog-content\") pod \"redhat-marketplace-fvw4b\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.519025 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-utilities\") pod \"redhat-marketplace-fvw4b\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.519094 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-catalog-content\") pod \"redhat-marketplace-fvw4b\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.543108 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtq8v\" (UniqueName: \"kubernetes.io/projected/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-kube-api-access-rtq8v\") pod \"redhat-marketplace-fvw4b\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.577280 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6pskk"] Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.578268 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.580829 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.592186 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6pskk"] Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.701232 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.721866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-utilities\") pod \"redhat-operators-6pskk\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.721920 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/282b8b99-95c9-45b5-9608-f57c52aae9e0-kube-api-access-qpmw2\") pod \"redhat-operators-6pskk\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.721952 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-catalog-content\") pod \"redhat-operators-6pskk\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.822664 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-utilities\") pod \"redhat-operators-6pskk\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.822959 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/282b8b99-95c9-45b5-9608-f57c52aae9e0-kube-api-access-qpmw2\") pod \"redhat-operators-6pskk\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.822995 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-catalog-content\") pod \"redhat-operators-6pskk\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.823217 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-utilities\") pod \"redhat-operators-6pskk\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.823317 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-catalog-content\") pod \"redhat-operators-6pskk\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.842030 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/282b8b99-95c9-45b5-9608-f57c52aae9e0-kube-api-access-qpmw2\") pod \"redhat-operators-6pskk\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.884572 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvw4b"] Nov 27 17:15:34 crc kubenswrapper[4809]: W1127 17:15:34.890890 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf04c3ed8_38c7_445a_b1b7_dfe646d8db71.slice/crio-b2ca0358e9473632226acfc01005a08dad1d4e0832a867eeff2ab5d61a9a1107 WatchSource:0}: Error finding container b2ca0358e9473632226acfc01005a08dad1d4e0832a867eeff2ab5d61a9a1107: Status 404 returned error can't find the container with id b2ca0358e9473632226acfc01005a08dad1d4e0832a867eeff2ab5d61a9a1107 Nov 27 17:15:34 crc kubenswrapper[4809]: I1127 17:15:34.897870 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:35 crc kubenswrapper[4809]: I1127 17:15:35.284573 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6pskk"] Nov 27 17:15:35 crc kubenswrapper[4809]: W1127 17:15:35.296934 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod282b8b99_95c9_45b5_9608_f57c52aae9e0.slice/crio-5eacc89d2dddc19cfa197f02ab79bf76c4b15e27e3e03c5cacee332cd36fe463 WatchSource:0}: Error finding container 5eacc89d2dddc19cfa197f02ab79bf76c4b15e27e3e03c5cacee332cd36fe463: Status 404 returned error can't find the container with id 5eacc89d2dddc19cfa197f02ab79bf76c4b15e27e3e03c5cacee332cd36fe463 Nov 27 17:15:35 crc kubenswrapper[4809]: I1127 17:15:35.336990 4809 generic.go:334] "Generic (PLEG): container finished" podID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerID="116203be209a4a802ed3f364fe944c6f2b96f645dbdf347e9a6018351dc4c4e6" exitCode=0 Nov 27 17:15:35 crc kubenswrapper[4809]: I1127 17:15:35.337071 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvw4b" event={"ID":"f04c3ed8-38c7-445a-b1b7-dfe646d8db71","Type":"ContainerDied","Data":"116203be209a4a802ed3f364fe944c6f2b96f645dbdf347e9a6018351dc4c4e6"} Nov 27 17:15:35 crc kubenswrapper[4809]: I1127 17:15:35.337103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvw4b" event={"ID":"f04c3ed8-38c7-445a-b1b7-dfe646d8db71","Type":"ContainerStarted","Data":"b2ca0358e9473632226acfc01005a08dad1d4e0832a867eeff2ab5d61a9a1107"} Nov 27 17:15:35 crc kubenswrapper[4809]: I1127 17:15:35.339126 4809 generic.go:334] "Generic (PLEG): container finished" podID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerID="fb3871a99fc82b18aca4345cb5e6d6a47649302dd57bf9409aa9dea99a8ee064" exitCode=0 Nov 27 17:15:35 crc kubenswrapper[4809]: I1127 17:15:35.339191 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgtgl" event={"ID":"ceaa3143-e2dc-40fd-8100-253811eeb357","Type":"ContainerDied","Data":"fb3871a99fc82b18aca4345cb5e6d6a47649302dd57bf9409aa9dea99a8ee064"} Nov 27 17:15:35 crc kubenswrapper[4809]: I1127 17:15:35.347553 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pskk" event={"ID":"282b8b99-95c9-45b5-9608-f57c52aae9e0","Type":"ContainerStarted","Data":"5eacc89d2dddc19cfa197f02ab79bf76c4b15e27e3e03c5cacee332cd36fe463"} Nov 27 17:15:35 crc kubenswrapper[4809]: I1127 17:15:35.350481 4809 generic.go:334] "Generic (PLEG): container finished" podID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerID="5c459bb53c76e8f422b63e4b290c124a84aa77ee269dca530d24c6f01585f3e2" exitCode=0 Nov 27 17:15:35 crc kubenswrapper[4809]: I1127 17:15:35.350510 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvpz" event={"ID":"57211ee0-a59b-4c8b-99d0-6854e00b9e53","Type":"ContainerDied","Data":"5c459bb53c76e8f422b63e4b290c124a84aa77ee269dca530d24c6f01585f3e2"} Nov 27 17:15:36 crc kubenswrapper[4809]: I1127 17:15:36.371924 4809 generic.go:334] "Generic (PLEG): container finished" podID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerID="100db1106b92fef5e567bd4b16ee21181be84b474fd88a2636a2baf6d1c6222e" exitCode=0 Nov 27 17:15:36 crc kubenswrapper[4809]: I1127 17:15:36.372033 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pskk" event={"ID":"282b8b99-95c9-45b5-9608-f57c52aae9e0","Type":"ContainerDied","Data":"100db1106b92fef5e567bd4b16ee21181be84b474fd88a2636a2baf6d1c6222e"} Nov 27 17:15:36 crc kubenswrapper[4809]: I1127 17:15:36.379766 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvpz" event={"ID":"57211ee0-a59b-4c8b-99d0-6854e00b9e53","Type":"ContainerStarted","Data":"e8cfe2308a0d685973241aac606be8f6c64aef23f0bc7403628ed57bd747e79d"} Nov 27 17:15:36 crc kubenswrapper[4809]: I1127 17:15:36.411914 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-swvpz" podStartSLOduration=1.8532110149999999 podStartE2EDuration="4.411890691s" podCreationTimestamp="2025-11-27 17:15:32 +0000 UTC" firstStartedPulling="2025-11-27 17:15:33.3229881 +0000 UTC m=+368.595445452" lastFinishedPulling="2025-11-27 17:15:35.881667776 +0000 UTC m=+371.154125128" observedRunningTime="2025-11-27 17:15:36.411364586 +0000 UTC m=+371.683821938" watchObservedRunningTime="2025-11-27 17:15:36.411890691 +0000 UTC m=+371.684348053" Nov 27 17:15:37 crc kubenswrapper[4809]: I1127 17:15:37.386647 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pskk" event={"ID":"282b8b99-95c9-45b5-9608-f57c52aae9e0","Type":"ContainerStarted","Data":"70cee96759554b5537cf9fdf0e16948f95bac4219620978c5c1fc4c2f289e86a"} Nov 27 17:15:37 crc kubenswrapper[4809]: I1127 17:15:37.389396 4809 generic.go:334] "Generic (PLEG): container finished" podID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerID="2ad80a4b4ea958ee1d3328c427830df599c322f09caa03c95cc74b8edfc4f62d" exitCode=0 Nov 27 17:15:37 crc kubenswrapper[4809]: I1127 17:15:37.389545 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvw4b" event={"ID":"f04c3ed8-38c7-445a-b1b7-dfe646d8db71","Type":"ContainerDied","Data":"2ad80a4b4ea958ee1d3328c427830df599c322f09caa03c95cc74b8edfc4f62d"} Nov 27 17:15:37 crc kubenswrapper[4809]: I1127 17:15:37.392086 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgtgl" event={"ID":"ceaa3143-e2dc-40fd-8100-253811eeb357","Type":"ContainerStarted","Data":"05278f23c41f669f29c49c75f01a1253460ed58a948c3c003cfa206717f4f18b"} Nov 27 17:15:37 crc kubenswrapper[4809]: I1127 17:15:37.433334 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pgtgl" podStartSLOduration=3.57119935 podStartE2EDuration="6.433318671s" podCreationTimestamp="2025-11-27 17:15:31 +0000 UTC" firstStartedPulling="2025-11-27 17:15:33.321316333 +0000 UTC m=+368.593773685" lastFinishedPulling="2025-11-27 17:15:36.183435654 +0000 UTC m=+371.455893006" observedRunningTime="2025-11-27 17:15:37.428422475 +0000 UTC m=+372.700879827" watchObservedRunningTime="2025-11-27 17:15:37.433318671 +0000 UTC m=+372.705776023" Nov 27 17:15:38 crc kubenswrapper[4809]: I1127 17:15:38.400081 4809 generic.go:334] "Generic (PLEG): container finished" podID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerID="70cee96759554b5537cf9fdf0e16948f95bac4219620978c5c1fc4c2f289e86a" exitCode=0 Nov 27 17:15:38 crc kubenswrapper[4809]: I1127 17:15:38.400140 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pskk" event={"ID":"282b8b99-95c9-45b5-9608-f57c52aae9e0","Type":"ContainerDied","Data":"70cee96759554b5537cf9fdf0e16948f95bac4219620978c5c1fc4c2f289e86a"} Nov 27 17:15:40 crc kubenswrapper[4809]: I1127 17:15:40.415370 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvw4b" event={"ID":"f04c3ed8-38c7-445a-b1b7-dfe646d8db71","Type":"ContainerStarted","Data":"c0a5608d1fba6506ac4040e74a1e5d6b93caa2adebd5e150366c2e1153d35527"} Nov 27 17:15:40 crc kubenswrapper[4809]: I1127 17:15:40.418445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pskk" event={"ID":"282b8b99-95c9-45b5-9608-f57c52aae9e0","Type":"ContainerStarted","Data":"818cf7664b2316410f662ce3719970dd8af11acd82bd6b3fa75392625b3b69c4"} Nov 27 17:15:40 crc kubenswrapper[4809]: I1127 17:15:40.444674 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fvw4b" podStartSLOduration=3.615122147 podStartE2EDuration="6.444648455s" podCreationTimestamp="2025-11-27 17:15:34 +0000 UTC" firstStartedPulling="2025-11-27 17:15:35.338405819 +0000 UTC m=+370.610863171" lastFinishedPulling="2025-11-27 17:15:38.167932127 +0000 UTC m=+373.440389479" observedRunningTime="2025-11-27 17:15:40.431831001 +0000 UTC m=+375.704288353" watchObservedRunningTime="2025-11-27 17:15:40.444648455 +0000 UTC m=+375.717105807" Nov 27 17:15:40 crc kubenswrapper[4809]: I1127 17:15:40.462028 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6pskk" podStartSLOduration=3.836785577 podStartE2EDuration="6.462008136s" podCreationTimestamp="2025-11-27 17:15:34 +0000 UTC" firstStartedPulling="2025-11-27 17:15:36.375340729 +0000 UTC m=+371.647798101" lastFinishedPulling="2025-11-27 17:15:39.000563298 +0000 UTC m=+374.273020660" observedRunningTime="2025-11-27 17:15:40.461927553 +0000 UTC m=+375.734384905" watchObservedRunningTime="2025-11-27 17:15:40.462008136 +0000 UTC m=+375.734465488" Nov 27 17:15:42 crc kubenswrapper[4809]: I1127 17:15:42.298946 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:42 crc kubenswrapper[4809]: I1127 17:15:42.299376 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:42 crc kubenswrapper[4809]: I1127 17:15:42.344138 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:42 crc kubenswrapper[4809]: I1127 17:15:42.468163 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:15:42 crc kubenswrapper[4809]: I1127 17:15:42.500825 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:42 crc kubenswrapper[4809]: I1127 17:15:42.500911 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:42 crc kubenswrapper[4809]: I1127 17:15:42.543433 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:43 crc kubenswrapper[4809]: I1127 17:15:43.471709 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:15:44 crc kubenswrapper[4809]: I1127 17:15:44.701922 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:44 crc kubenswrapper[4809]: I1127 17:15:44.701986 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:44 crc kubenswrapper[4809]: I1127 17:15:44.745616 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:44 crc kubenswrapper[4809]: I1127 17:15:44.897994 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:44 crc kubenswrapper[4809]: I1127 17:15:44.898040 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:44 crc kubenswrapper[4809]: I1127 17:15:44.941402 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:45 crc kubenswrapper[4809]: I1127 17:15:45.480515 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:15:45 crc kubenswrapper[4809]: I1127 17:15:45.481526 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:15:55 crc kubenswrapper[4809]: I1127 17:15:55.779956 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:15:55 crc kubenswrapper[4809]: I1127 17:15:55.780930 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:15:59 crc kubenswrapper[4809]: I1127 17:15:59.550570 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" podUID="480a6028-0611-40c1-8204-0e2e37608800" containerName="registry" containerID="cri-o://49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349" gracePeriod=30 Nov 27 17:15:59 crc kubenswrapper[4809]: I1127 17:15:59.912679 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.081195 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"480a6028-0611-40c1-8204-0e2e37608800\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.081280 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/480a6028-0611-40c1-8204-0e2e37608800-ca-trust-extracted\") pod \"480a6028-0611-40c1-8204-0e2e37608800\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.081343 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-registry-certificates\") pod \"480a6028-0611-40c1-8204-0e2e37608800\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.081364 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x47mb\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-kube-api-access-x47mb\") pod \"480a6028-0611-40c1-8204-0e2e37608800\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.081399 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-bound-sa-token\") pod \"480a6028-0611-40c1-8204-0e2e37608800\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.081447 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/480a6028-0611-40c1-8204-0e2e37608800-installation-pull-secrets\") pod \"480a6028-0611-40c1-8204-0e2e37608800\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.081487 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-trusted-ca\") pod \"480a6028-0611-40c1-8204-0e2e37608800\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.081513 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-registry-tls\") pod \"480a6028-0611-40c1-8204-0e2e37608800\" (UID: \"480a6028-0611-40c1-8204-0e2e37608800\") " Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.082387 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "480a6028-0611-40c1-8204-0e2e37608800" (UID: "480a6028-0611-40c1-8204-0e2e37608800"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.082450 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "480a6028-0611-40c1-8204-0e2e37608800" (UID: "480a6028-0611-40c1-8204-0e2e37608800"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.089709 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a6028-0611-40c1-8204-0e2e37608800-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "480a6028-0611-40c1-8204-0e2e37608800" (UID: "480a6028-0611-40c1-8204-0e2e37608800"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.090474 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "480a6028-0611-40c1-8204-0e2e37608800" (UID: "480a6028-0611-40c1-8204-0e2e37608800"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.091109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "480a6028-0611-40c1-8204-0e2e37608800" (UID: "480a6028-0611-40c1-8204-0e2e37608800"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.092058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-kube-api-access-x47mb" (OuterVolumeSpecName: "kube-api-access-x47mb") pod "480a6028-0611-40c1-8204-0e2e37608800" (UID: "480a6028-0611-40c1-8204-0e2e37608800"). InnerVolumeSpecName "kube-api-access-x47mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.092353 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "480a6028-0611-40c1-8204-0e2e37608800" (UID: "480a6028-0611-40c1-8204-0e2e37608800"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.099168 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/480a6028-0611-40c1-8204-0e2e37608800-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "480a6028-0611-40c1-8204-0e2e37608800" (UID: "480a6028-0611-40c1-8204-0e2e37608800"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.182987 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.183030 4809 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.183041 4809 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/480a6028-0611-40c1-8204-0e2e37608800-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.183050 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x47mb\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-kube-api-access-x47mb\") on node \"crc\" DevicePath \"\"" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.183061 4809 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/480a6028-0611-40c1-8204-0e2e37608800-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.183071 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/480a6028-0611-40c1-8204-0e2e37608800-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.183081 4809 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/480a6028-0611-40c1-8204-0e2e37608800-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.528762 4809 generic.go:334] "Generic (PLEG): container finished" podID="480a6028-0611-40c1-8204-0e2e37608800" containerID="49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349" exitCode=0 Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.528841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" event={"ID":"480a6028-0611-40c1-8204-0e2e37608800","Type":"ContainerDied","Data":"49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349"} Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.528890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" event={"ID":"480a6028-0611-40c1-8204-0e2e37608800","Type":"ContainerDied","Data":"6cb3bf7ce5beb021d7744a931799e9f9397f8992cadaf92bd9ce73ab8acbb96f"} Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.528912 4809 scope.go:117] "RemoveContainer" containerID="49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.528916 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dj24l" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.558360 4809 scope.go:117] "RemoveContainer" containerID="49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349" Nov 27 17:16:00 crc kubenswrapper[4809]: E1127 17:16:00.559378 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349\": container with ID starting with 49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349 not found: ID does not exist" containerID="49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.559519 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349"} err="failed to get container status \"49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349\": rpc error: code = NotFound desc = could not find container \"49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349\": container with ID starting with 49147212d3f5bd64c79c8c591cec2f2571168c249ed4c4c1108218c0c0b2d349 not found: ID does not exist" Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.572517 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dj24l"] Nov 27 17:16:00 crc kubenswrapper[4809]: I1127 17:16:00.580184 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dj24l"] Nov 27 17:16:01 crc kubenswrapper[4809]: I1127 17:16:01.470710 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480a6028-0611-40c1-8204-0e2e37608800" path="/var/lib/kubelet/pods/480a6028-0611-40c1-8204-0e2e37608800/volumes" Nov 27 17:16:25 crc kubenswrapper[4809]: I1127 17:16:25.779537 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:16:25 crc kubenswrapper[4809]: I1127 17:16:25.780087 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:16:25 crc kubenswrapper[4809]: I1127 17:16:25.780139 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:16:25 crc kubenswrapper[4809]: I1127 17:16:25.780823 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc8b858c869d6867bc225e685a98fd0cfd9e143945af5b8b25449257c7e1fd3a"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:16:25 crc kubenswrapper[4809]: I1127 17:16:25.780883 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://bc8b858c869d6867bc225e685a98fd0cfd9e143945af5b8b25449257c7e1fd3a" gracePeriod=600 Nov 27 17:16:26 crc kubenswrapper[4809]: I1127 17:16:26.686613 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="bc8b858c869d6867bc225e685a98fd0cfd9e143945af5b8b25449257c7e1fd3a" exitCode=0 Nov 27 17:16:26 crc kubenswrapper[4809]: I1127 17:16:26.686691 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"bc8b858c869d6867bc225e685a98fd0cfd9e143945af5b8b25449257c7e1fd3a"} Nov 27 17:16:26 crc kubenswrapper[4809]: I1127 17:16:26.687289 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"cb5f4afccabcb96c09b1f91bcfcd46aff8f41c149de91ed16add7b6ee4b7e98c"} Nov 27 17:16:26 crc kubenswrapper[4809]: I1127 17:16:26.687317 4809 scope.go:117] "RemoveContainer" containerID="b8e58cab8e57595232f47dfe66ff47e1048c4840a47afb755d7ef49b7976fb75" Nov 27 17:18:55 crc kubenswrapper[4809]: I1127 17:18:55.779978 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:18:55 crc kubenswrapper[4809]: I1127 17:18:55.780523 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:19:25 crc kubenswrapper[4809]: I1127 17:19:25.778997 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:19:25 crc kubenswrapper[4809]: I1127 17:19:25.779641 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:19:55 crc kubenswrapper[4809]: I1127 17:19:55.779126 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:19:55 crc kubenswrapper[4809]: I1127 17:19:55.779760 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:19:55 crc kubenswrapper[4809]: I1127 17:19:55.779821 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:19:55 crc kubenswrapper[4809]: I1127 17:19:55.780444 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb5f4afccabcb96c09b1f91bcfcd46aff8f41c149de91ed16add7b6ee4b7e98c"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:19:55 crc kubenswrapper[4809]: I1127 17:19:55.780506 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://cb5f4afccabcb96c09b1f91bcfcd46aff8f41c149de91ed16add7b6ee4b7e98c" gracePeriod=600 Nov 27 17:19:55 crc kubenswrapper[4809]: I1127 17:19:55.909844 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="cb5f4afccabcb96c09b1f91bcfcd46aff8f41c149de91ed16add7b6ee4b7e98c" exitCode=0 Nov 27 17:19:55 crc kubenswrapper[4809]: I1127 17:19:55.909944 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"cb5f4afccabcb96c09b1f91bcfcd46aff8f41c149de91ed16add7b6ee4b7e98c"} Nov 27 17:19:55 crc kubenswrapper[4809]: I1127 17:19:55.910019 4809 scope.go:117] "RemoveContainer" containerID="bc8b858c869d6867bc225e685a98fd0cfd9e143945af5b8b25449257c7e1fd3a" Nov 27 17:19:56 crc kubenswrapper[4809]: I1127 17:19:56.918345 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"4c5bbc3124a8972c9912c532472d23e13e06555c5f6e029a3731193652f77944"} Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.947195 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rczjv"] Nov 27 17:21:39 crc kubenswrapper[4809]: E1127 17:21:39.948298 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480a6028-0611-40c1-8204-0e2e37608800" containerName="registry" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.948314 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="480a6028-0611-40c1-8204-0e2e37608800" containerName="registry" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.948418 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="480a6028-0611-40c1-8204-0e2e37608800" containerName="registry" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.948911 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rczjv" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.958544 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7hzbp"] Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.959516 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7hzbp" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.964016 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-bntnl" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.966935 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.981313 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.983238 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-fg5ng" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.985935 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rczjv"] Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.989486 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7hzbp"] Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.992705 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-tkpwp"] Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.993478 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" Nov 27 17:21:39 crc kubenswrapper[4809]: I1127 17:21:39.996475 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-q6dm4" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.003112 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-tkpwp"] Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.109850 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7hc4\" (UniqueName: \"kubernetes.io/projected/79bf3153-df84-43fd-baa1-731fbc7f7873-kube-api-access-d7hc4\") pod \"cert-manager-webhook-5655c58dd6-tkpwp\" (UID: \"79bf3153-df84-43fd-baa1-731fbc7f7873\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.109899 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rfp4\" (UniqueName: \"kubernetes.io/projected/04d521a8-9cbf-4ca8-b7a8-035504c16c17-kube-api-access-8rfp4\") pod \"cert-manager-cainjector-7f985d654d-rczjv\" (UID: \"04d521a8-9cbf-4ca8-b7a8-035504c16c17\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rczjv" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.110014 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwcx7\" (UniqueName: \"kubernetes.io/projected/df88093a-9522-45ca-913a-18a62128af63-kube-api-access-vwcx7\") pod \"cert-manager-5b446d88c5-7hzbp\" (UID: \"df88093a-9522-45ca-913a-18a62128af63\") " pod="cert-manager/cert-manager-5b446d88c5-7hzbp" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.211530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwcx7\" (UniqueName: \"kubernetes.io/projected/df88093a-9522-45ca-913a-18a62128af63-kube-api-access-vwcx7\") pod \"cert-manager-5b446d88c5-7hzbp\" (UID: \"df88093a-9522-45ca-913a-18a62128af63\") " pod="cert-manager/cert-manager-5b446d88c5-7hzbp" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.211691 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7hc4\" (UniqueName: \"kubernetes.io/projected/79bf3153-df84-43fd-baa1-731fbc7f7873-kube-api-access-d7hc4\") pod \"cert-manager-webhook-5655c58dd6-tkpwp\" (UID: \"79bf3153-df84-43fd-baa1-731fbc7f7873\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.211788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rfp4\" (UniqueName: \"kubernetes.io/projected/04d521a8-9cbf-4ca8-b7a8-035504c16c17-kube-api-access-8rfp4\") pod \"cert-manager-cainjector-7f985d654d-rczjv\" (UID: \"04d521a8-9cbf-4ca8-b7a8-035504c16c17\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rczjv" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.237293 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7hc4\" (UniqueName: \"kubernetes.io/projected/79bf3153-df84-43fd-baa1-731fbc7f7873-kube-api-access-d7hc4\") pod \"cert-manager-webhook-5655c58dd6-tkpwp\" (UID: \"79bf3153-df84-43fd-baa1-731fbc7f7873\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.237360 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rfp4\" (UniqueName: \"kubernetes.io/projected/04d521a8-9cbf-4ca8-b7a8-035504c16c17-kube-api-access-8rfp4\") pod \"cert-manager-cainjector-7f985d654d-rczjv\" (UID: \"04d521a8-9cbf-4ca8-b7a8-035504c16c17\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rczjv" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.238695 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwcx7\" (UniqueName: \"kubernetes.io/projected/df88093a-9522-45ca-913a-18a62128af63-kube-api-access-vwcx7\") pod \"cert-manager-5b446d88c5-7hzbp\" (UID: \"df88093a-9522-45ca-913a-18a62128af63\") " pod="cert-manager/cert-manager-5b446d88c5-7hzbp" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.267651 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rczjv" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.279125 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7hzbp" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.306916 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.560287 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-tkpwp"] Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.570436 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.710756 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rczjv"] Nov 27 17:21:40 crc kubenswrapper[4809]: W1127 17:21:40.718176 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04d521a8_9cbf_4ca8_b7a8_035504c16c17.slice/crio-8192076119cfc2687a590650fb6e02901d1e07031a9c8422ab909a65ec2a6d41 WatchSource:0}: Error finding container 8192076119cfc2687a590650fb6e02901d1e07031a9c8422ab909a65ec2a6d41: Status 404 returned error can't find the container with id 8192076119cfc2687a590650fb6e02901d1e07031a9c8422ab909a65ec2a6d41 Nov 27 17:21:40 crc kubenswrapper[4809]: I1127 17:21:40.723074 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7hzbp"] Nov 27 17:21:40 crc kubenswrapper[4809]: W1127 17:21:40.724496 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf88093a_9522_45ca_913a_18a62128af63.slice/crio-b637daf9a7d750c22bc53796d76ee410d66a59dd2ac2d47d966c7f03f467d3c6 WatchSource:0}: Error finding container b637daf9a7d750c22bc53796d76ee410d66a59dd2ac2d47d966c7f03f467d3c6: Status 404 returned error can't find the container with id b637daf9a7d750c22bc53796d76ee410d66a59dd2ac2d47d966c7f03f467d3c6 Nov 27 17:21:41 crc kubenswrapper[4809]: I1127 17:21:41.547381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rczjv" event={"ID":"04d521a8-9cbf-4ca8-b7a8-035504c16c17","Type":"ContainerStarted","Data":"8192076119cfc2687a590650fb6e02901d1e07031a9c8422ab909a65ec2a6d41"} Nov 27 17:21:41 crc kubenswrapper[4809]: I1127 17:21:41.554069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7hzbp" event={"ID":"df88093a-9522-45ca-913a-18a62128af63","Type":"ContainerStarted","Data":"b637daf9a7d750c22bc53796d76ee410d66a59dd2ac2d47d966c7f03f467d3c6"} Nov 27 17:21:41 crc kubenswrapper[4809]: I1127 17:21:41.555866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" event={"ID":"79bf3153-df84-43fd-baa1-731fbc7f7873","Type":"ContainerStarted","Data":"f13ef17f78f5e50bf335a6c4f40afb76e4fcfc449ba2b6f1e9be3ef5f6f575d3"} Nov 27 17:21:44 crc kubenswrapper[4809]: I1127 17:21:44.574660 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rczjv" event={"ID":"04d521a8-9cbf-4ca8-b7a8-035504c16c17","Type":"ContainerStarted","Data":"0dd4f17f41a0e29520410f8056e0586fb4016fdf2ff933607241a08a5f139c42"} Nov 27 17:21:44 crc kubenswrapper[4809]: I1127 17:21:44.578509 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7hzbp" event={"ID":"df88093a-9522-45ca-913a-18a62128af63","Type":"ContainerStarted","Data":"9ad2bc6864908a52fa00ff7a88399b18d5cf0f66ba01fa45311c1e4a590a32dd"} Nov 27 17:21:44 crc kubenswrapper[4809]: I1127 17:21:44.582249 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" event={"ID":"79bf3153-df84-43fd-baa1-731fbc7f7873","Type":"ContainerStarted","Data":"135436795af2b0a47250689a07c0a6053f43b977eb15978e0cf41d86878344e9"} Nov 27 17:21:44 crc kubenswrapper[4809]: I1127 17:21:44.582396 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" Nov 27 17:21:44 crc kubenswrapper[4809]: I1127 17:21:44.590522 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-rczjv" podStartSLOduration=2.757938044 podStartE2EDuration="5.590501179s" podCreationTimestamp="2025-11-27 17:21:39 +0000 UTC" firstStartedPulling="2025-11-27 17:21:40.721097942 +0000 UTC m=+735.993555294" lastFinishedPulling="2025-11-27 17:21:43.553661077 +0000 UTC m=+738.826118429" observedRunningTime="2025-11-27 17:21:44.586131458 +0000 UTC m=+739.858588810" watchObservedRunningTime="2025-11-27 17:21:44.590501179 +0000 UTC m=+739.862958521" Nov 27 17:21:44 crc kubenswrapper[4809]: I1127 17:21:44.602588 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" podStartSLOduration=2.610698616 podStartE2EDuration="5.6025509s" podCreationTimestamp="2025-11-27 17:21:39 +0000 UTC" firstStartedPulling="2025-11-27 17:21:40.570210584 +0000 UTC m=+735.842667936" lastFinishedPulling="2025-11-27 17:21:43.562062858 +0000 UTC m=+738.834520220" observedRunningTime="2025-11-27 17:21:44.599173217 +0000 UTC m=+739.871630569" watchObservedRunningTime="2025-11-27 17:21:44.6025509 +0000 UTC m=+739.875008242" Nov 27 17:21:44 crc kubenswrapper[4809]: I1127 17:21:44.617388 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-7hzbp" podStartSLOduration=2.780949155 podStartE2EDuration="5.617366337s" podCreationTimestamp="2025-11-27 17:21:39 +0000 UTC" firstStartedPulling="2025-11-27 17:21:40.728753861 +0000 UTC m=+736.001211203" lastFinishedPulling="2025-11-27 17:21:43.565171043 +0000 UTC m=+738.837628385" observedRunningTime="2025-11-27 17:21:44.61422423 +0000 UTC m=+739.886681582" watchObservedRunningTime="2025-11-27 17:21:44.617366337 +0000 UTC m=+739.889823679" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.311704 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-tkpwp" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.488910 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-m5b98"] Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.489344 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovn-controller" containerID="cri-o://a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a" gracePeriod=30 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.489452 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="northd" containerID="cri-o://31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45" gracePeriod=30 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.489539 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9" gracePeriod=30 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.489506 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kube-rbac-proxy-node" containerID="cri-o://ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc" gracePeriod=30 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.489510 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovn-acl-logging" containerID="cri-o://b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3" gracePeriod=30 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.489658 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="sbdb" containerID="cri-o://0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8" gracePeriod=30 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.489668 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="nbdb" containerID="cri-o://d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1" gracePeriod=30 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.518807 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" containerID="cri-o://d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b" gracePeriod=30 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.625819 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd8px_8741a6e8-a7f2-40ef-b25b-10e518345478/kube-multus/2.log" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.627065 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd8px_8741a6e8-a7f2-40ef-b25b-10e518345478/kube-multus/1.log" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.627119 4809 generic.go:334] "Generic (PLEG): container finished" podID="8741a6e8-a7f2-40ef-b25b-10e518345478" containerID="090893af68773017e5a04393586f2d8375589b6156f7fca536afd90c178c9f26" exitCode=2 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.627198 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd8px" event={"ID":"8741a6e8-a7f2-40ef-b25b-10e518345478","Type":"ContainerDied","Data":"090893af68773017e5a04393586f2d8375589b6156f7fca536afd90c178c9f26"} Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.627258 4809 scope.go:117] "RemoveContainer" containerID="aa298fe25c9ddb39a55a73d0dcc0418247d88f1b6c3d9e119dca173543492156" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.628164 4809 scope.go:117] "RemoveContainer" containerID="090893af68773017e5a04393586f2d8375589b6156f7fca536afd90c178c9f26" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.631132 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/3.log" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.636101 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovn-acl-logging/0.log" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.636728 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovn-controller/0.log" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.637507 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9" exitCode=0 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.637562 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc" exitCode=0 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.637578 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3" exitCode=143 Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.637584 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9"} Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.637647 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc"} Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.637665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3"} Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.858889 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/3.log" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.862362 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovn-acl-logging/0.log" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.863076 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovn-controller/0.log" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.863632 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922262 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xtll8"] Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922538 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovn-acl-logging" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922561 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovn-acl-logging" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922572 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kubecfg-setup" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922579 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kubecfg-setup" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922592 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922599 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922608 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="nbdb" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922614 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="nbdb" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922623 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922629 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922639 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922644 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922653 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922661 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922682 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="sbdb" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922689 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="sbdb" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922697 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovn-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922703 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovn-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922712 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="northd" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922719 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="northd" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.922729 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kube-rbac-proxy-node" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922738 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kube-rbac-proxy-node" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922855 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kube-rbac-proxy-node" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922867 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922875 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="nbdb" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922881 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922889 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovn-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922897 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922903 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922911 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922917 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovn-acl-logging" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922924 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="northd" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922934 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="sbdb" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.922942 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.923038 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.923047 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: E1127 17:21:50.923056 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.923062 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b4318a-f089-451d-8a16-97de26acce28" containerName="ovnkube-controller" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.924900 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.958912 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-var-lib-openvswitch\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.958996 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-node-log\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959065 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-script-lib\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959072 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959120 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-node-log" (OuterVolumeSpecName: "node-log") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959154 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959095 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-netns\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959232 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-kubelet\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959269 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-bin\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959295 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959318 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-systemd\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959332 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959351 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-netd\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959356 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959380 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-env-overrides\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959411 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959425 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959557 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959624 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-var-lib-openvswitch\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959653 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/49b29b48-c4c5-403b-ba79-21c18dda8430-env-overrides\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-etc-openvswitch\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/49b29b48-c4c5-403b-ba79-21c18dda8430-ovnkube-script-lib\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959741 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-kubelet\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959795 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959810 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-run-openvswitch\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959835 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-run-ovn\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.959948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/49b29b48-c4c5-403b-ba79-21c18dda8430-ovnkube-config\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960010 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-cni-netd\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960034 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-cni-bin\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960052 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/49b29b48-c4c5-403b-ba79-21c18dda8430-ovn-node-metrics-cert\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960079 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6h4z\" (UniqueName: \"kubernetes.io/projected/49b29b48-c4c5-403b-ba79-21c18dda8430-kube-api-access-q6h4z\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960108 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-log-socket\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960156 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-slash\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960195 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-run-systemd\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960215 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-systemd-units\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-run-netns\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960255 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-node-log\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-run-ovn-kubernetes\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960385 4809 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960412 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960434 4809 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960451 4809 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-node-log\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960463 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960478 4809 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960491 4809 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960502 4809 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.960518 4809 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:50 crc kubenswrapper[4809]: I1127 17:21:50.975158 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.060807 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52b7c\" (UniqueName: \"kubernetes.io/projected/66b4318a-f089-451d-8a16-97de26acce28-kube-api-access-52b7c\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.060871 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66b4318a-f089-451d-8a16-97de26acce28-ovn-node-metrics-cert\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.060900 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-systemd-units\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.060920 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-etc-openvswitch\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.060966 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-ovn-kubernetes\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.060990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-openvswitch\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061022 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-config\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061055 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-ovn\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061078 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-slash\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061109 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-log-socket\") pod \"66b4318a-f089-451d-8a16-97de26acce28\" (UID: \"66b4318a-f089-451d-8a16-97de26acce28\") " Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061210 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-run-openvswitch\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061238 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-run-ovn\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/49b29b48-c4c5-403b-ba79-21c18dda8430-ovnkube-config\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061311 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-cni-netd\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061333 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-cni-bin\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061319 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061347 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061438 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061357 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/49b29b48-c4c5-403b-ba79-21c18dda8430-ovn-node-metrics-cert\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061561 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061599 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-slash" (OuterVolumeSpecName: "host-slash") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061631 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-log-socket" (OuterVolumeSpecName: "log-socket") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061855 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-run-openvswitch\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061921 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.061969 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-cni-netd\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062007 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-run-ovn\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062038 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-cni-bin\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062133 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6h4z\" (UniqueName: \"kubernetes.io/projected/49b29b48-c4c5-403b-ba79-21c18dda8430-kube-api-access-q6h4z\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062160 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-log-socket\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062184 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-slash\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062206 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-run-systemd\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-systemd-units\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-run-netns\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062277 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-node-log\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062301 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-run-ovn-kubernetes\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/49b29b48-c4c5-403b-ba79-21c18dda8430-env-overrides\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062363 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-var-lib-openvswitch\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062394 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-etc-openvswitch\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062421 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/49b29b48-c4c5-403b-ba79-21c18dda8430-ovnkube-script-lib\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-kubelet\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062527 4809 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-log-socket\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062545 4809 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062558 4809 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062572 4809 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062584 4809 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062597 4809 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062610 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66b4318a-f089-451d-8a16-97de26acce28-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062624 4809 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062637 4809 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/66b4318a-f089-451d-8a16-97de26acce28-host-slash\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-node-log\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.062687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-run-ovn-kubernetes\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063005 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/49b29b48-c4c5-403b-ba79-21c18dda8430-ovnkube-config\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063072 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-log-socket\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063095 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-run-systemd\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063144 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-systemd-units\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063148 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-slash\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063182 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-run-netns\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063230 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-var-lib-openvswitch\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063276 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-etc-openvswitch\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063285 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/49b29b48-c4c5-403b-ba79-21c18dda8430-env-overrides\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063321 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.063341 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/49b29b48-c4c5-403b-ba79-21c18dda8430-host-kubelet\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.064186 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/49b29b48-c4c5-403b-ba79-21c18dda8430-ovnkube-script-lib\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.066903 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/49b29b48-c4c5-403b-ba79-21c18dda8430-ovn-node-metrics-cert\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.067197 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b4318a-f089-451d-8a16-97de26acce28-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.072611 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b4318a-f089-451d-8a16-97de26acce28-kube-api-access-52b7c" (OuterVolumeSpecName: "kube-api-access-52b7c") pod "66b4318a-f089-451d-8a16-97de26acce28" (UID: "66b4318a-f089-451d-8a16-97de26acce28"). InnerVolumeSpecName "kube-api-access-52b7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.086731 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6h4z\" (UniqueName: \"kubernetes.io/projected/49b29b48-c4c5-403b-ba79-21c18dda8430-kube-api-access-q6h4z\") pod \"ovnkube-node-xtll8\" (UID: \"49b29b48-c4c5-403b-ba79-21c18dda8430\") " pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.164046 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52b7c\" (UniqueName: \"kubernetes.io/projected/66b4318a-f089-451d-8a16-97de26acce28-kube-api-access-52b7c\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.164099 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66b4318a-f089-451d-8a16-97de26acce28-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.237647 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:51 crc kubenswrapper[4809]: W1127 17:21:51.263859 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49b29b48_c4c5_403b_ba79_21c18dda8430.slice/crio-d93d5ebfefe86c5361e82d350f0772181b174a0a21d3bf7461ee5d32978b3254 WatchSource:0}: Error finding container d93d5ebfefe86c5361e82d350f0772181b174a0a21d3bf7461ee5d32978b3254: Status 404 returned error can't find the container with id d93d5ebfefe86c5361e82d350f0772181b174a0a21d3bf7461ee5d32978b3254 Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.646609 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd8px_8741a6e8-a7f2-40ef-b25b-10e518345478/kube-multus/2.log" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.646813 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd8px" event={"ID":"8741a6e8-a7f2-40ef-b25b-10e518345478","Type":"ContainerStarted","Data":"c62e7603afdcd3546a9297f03df4c8d4c59e8be486d2f0ee7e890cbf0ccacd1d"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.650710 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovnkube-controller/3.log" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.653387 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovn-acl-logging/0.log" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.653798 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m5b98_66b4318a-f089-451d-8a16-97de26acce28/ovn-controller/0.log" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654071 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b" exitCode=0 Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654095 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8" exitCode=0 Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654102 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1" exitCode=0 Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654108 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45" exitCode=0 Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654116 4809 generic.go:334] "Generic (PLEG): container finished" podID="66b4318a-f089-451d-8a16-97de26acce28" containerID="a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a" exitCode=143 Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654159 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654187 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654197 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654207 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654216 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654227 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" event={"ID":"66b4318a-f089-451d-8a16-97de26acce28","Type":"ContainerDied","Data":"f5e01e2601c4021d050c567e3eaa7fbaff2144102276775b24d497d2ba951052"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654237 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654247 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654253 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654258 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654263 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654268 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654274 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654281 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654286 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654299 4809 scope.go:117] "RemoveContainer" containerID="d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.654406 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-m5b98" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.656379 4809 generic.go:334] "Generic (PLEG): container finished" podID="49b29b48-c4c5-403b-ba79-21c18dda8430" containerID="fe72348f2295ed2c579340dfd096a7b01ea9a0712cf2cf0a99c5d1c5c4a5d6bc" exitCode=0 Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.656433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerDied","Data":"fe72348f2295ed2c579340dfd096a7b01ea9a0712cf2cf0a99c5d1c5c4a5d6bc"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.656482 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerStarted","Data":"d93d5ebfefe86c5361e82d350f0772181b174a0a21d3bf7461ee5d32978b3254"} Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.687380 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.726859 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-m5b98"] Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.738855 4809 scope.go:117] "RemoveContainer" containerID="0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.740233 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-m5b98"] Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.778687 4809 scope.go:117] "RemoveContainer" containerID="d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.800022 4809 scope.go:117] "RemoveContainer" containerID="31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.818045 4809 scope.go:117] "RemoveContainer" containerID="7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.838374 4809 scope.go:117] "RemoveContainer" containerID="ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.857253 4809 scope.go:117] "RemoveContainer" containerID="b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.888194 4809 scope.go:117] "RemoveContainer" containerID="a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.921456 4809 scope.go:117] "RemoveContainer" containerID="dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.940311 4809 scope.go:117] "RemoveContainer" containerID="d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.940834 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": container with ID starting with d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b not found: ID does not exist" containerID="d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.940881 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b"} err="failed to get container status \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": rpc error: code = NotFound desc = could not find container \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": container with ID starting with d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.940912 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.941452 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\": container with ID starting with 2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d not found: ID does not exist" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.941525 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d"} err="failed to get container status \"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\": rpc error: code = NotFound desc = could not find container \"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\": container with ID starting with 2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.941575 4809 scope.go:117] "RemoveContainer" containerID="0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.942260 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\": container with ID starting with 0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8 not found: ID does not exist" containerID="0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.942324 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8"} err="failed to get container status \"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\": rpc error: code = NotFound desc = could not find container \"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\": container with ID starting with 0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.942360 4809 scope.go:117] "RemoveContainer" containerID="d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.943285 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\": container with ID starting with d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1 not found: ID does not exist" containerID="d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.943329 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1"} err="failed to get container status \"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\": rpc error: code = NotFound desc = could not find container \"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\": container with ID starting with d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.943358 4809 scope.go:117] "RemoveContainer" containerID="31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.944158 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\": container with ID starting with 31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45 not found: ID does not exist" containerID="31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.944189 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45"} err="failed to get container status \"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\": rpc error: code = NotFound desc = could not find container \"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\": container with ID starting with 31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.944207 4809 scope.go:117] "RemoveContainer" containerID="7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.944546 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\": container with ID starting with 7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9 not found: ID does not exist" containerID="7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.944569 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9"} err="failed to get container status \"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\": rpc error: code = NotFound desc = could not find container \"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\": container with ID starting with 7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.944582 4809 scope.go:117] "RemoveContainer" containerID="ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.945125 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\": container with ID starting with ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc not found: ID does not exist" containerID="ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.945174 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc"} err="failed to get container status \"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\": rpc error: code = NotFound desc = could not find container \"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\": container with ID starting with ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.945214 4809 scope.go:117] "RemoveContainer" containerID="b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.945552 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\": container with ID starting with b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3 not found: ID does not exist" containerID="b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.945590 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3"} err="failed to get container status \"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\": rpc error: code = NotFound desc = could not find container \"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\": container with ID starting with b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.945618 4809 scope.go:117] "RemoveContainer" containerID="a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.946038 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\": container with ID starting with a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a not found: ID does not exist" containerID="a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.946065 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a"} err="failed to get container status \"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\": rpc error: code = NotFound desc = could not find container \"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\": container with ID starting with a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.946084 4809 scope.go:117] "RemoveContainer" containerID="dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f" Nov 27 17:21:51 crc kubenswrapper[4809]: E1127 17:21:51.946403 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\": container with ID starting with dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f not found: ID does not exist" containerID="dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.946437 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f"} err="failed to get container status \"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\": rpc error: code = NotFound desc = could not find container \"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\": container with ID starting with dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.946463 4809 scope.go:117] "RemoveContainer" containerID="d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.946713 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b"} err="failed to get container status \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": rpc error: code = NotFound desc = could not find container \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": container with ID starting with d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.946732 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.947078 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d"} err="failed to get container status \"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\": rpc error: code = NotFound desc = could not find container \"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\": container with ID starting with 2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.947101 4809 scope.go:117] "RemoveContainer" containerID="0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.947398 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8"} err="failed to get container status \"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\": rpc error: code = NotFound desc = could not find container \"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\": container with ID starting with 0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.947434 4809 scope.go:117] "RemoveContainer" containerID="d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.947796 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1"} err="failed to get container status \"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\": rpc error: code = NotFound desc = could not find container \"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\": container with ID starting with d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.947821 4809 scope.go:117] "RemoveContainer" containerID="31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.948065 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45"} err="failed to get container status \"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\": rpc error: code = NotFound desc = could not find container \"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\": container with ID starting with 31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.948088 4809 scope.go:117] "RemoveContainer" containerID="7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.948354 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9"} err="failed to get container status \"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\": rpc error: code = NotFound desc = could not find container \"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\": container with ID starting with 7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.948381 4809 scope.go:117] "RemoveContainer" containerID="ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.948619 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc"} err="failed to get container status \"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\": rpc error: code = NotFound desc = could not find container \"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\": container with ID starting with ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.948646 4809 scope.go:117] "RemoveContainer" containerID="b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.948898 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3"} err="failed to get container status \"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\": rpc error: code = NotFound desc = could not find container \"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\": container with ID starting with b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.948929 4809 scope.go:117] "RemoveContainer" containerID="a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.949144 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a"} err="failed to get container status \"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\": rpc error: code = NotFound desc = could not find container \"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\": container with ID starting with a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.949165 4809 scope.go:117] "RemoveContainer" containerID="dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.949379 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f"} err="failed to get container status \"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\": rpc error: code = NotFound desc = could not find container \"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\": container with ID starting with dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.949401 4809 scope.go:117] "RemoveContainer" containerID="d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.949714 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b"} err="failed to get container status \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": rpc error: code = NotFound desc = could not find container \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": container with ID starting with d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.949741 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.949975 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d"} err="failed to get container status \"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\": rpc error: code = NotFound desc = could not find container \"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\": container with ID starting with 2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.949999 4809 scope.go:117] "RemoveContainer" containerID="0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.950239 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8"} err="failed to get container status \"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\": rpc error: code = NotFound desc = could not find container \"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\": container with ID starting with 0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.950263 4809 scope.go:117] "RemoveContainer" containerID="d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.950472 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1"} err="failed to get container status \"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\": rpc error: code = NotFound desc = could not find container \"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\": container with ID starting with d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.950499 4809 scope.go:117] "RemoveContainer" containerID="31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.950721 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45"} err="failed to get container status \"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\": rpc error: code = NotFound desc = could not find container \"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\": container with ID starting with 31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.950746 4809 scope.go:117] "RemoveContainer" containerID="7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.951031 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9"} err="failed to get container status \"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\": rpc error: code = NotFound desc = could not find container \"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\": container with ID starting with 7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.951054 4809 scope.go:117] "RemoveContainer" containerID="ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.951303 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc"} err="failed to get container status \"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\": rpc error: code = NotFound desc = could not find container \"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\": container with ID starting with ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.951327 4809 scope.go:117] "RemoveContainer" containerID="b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.951557 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3"} err="failed to get container status \"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\": rpc error: code = NotFound desc = could not find container \"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\": container with ID starting with b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.951578 4809 scope.go:117] "RemoveContainer" containerID="a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.951780 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a"} err="failed to get container status \"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\": rpc error: code = NotFound desc = could not find container \"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\": container with ID starting with a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.951802 4809 scope.go:117] "RemoveContainer" containerID="dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.952028 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f"} err="failed to get container status \"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\": rpc error: code = NotFound desc = could not find container \"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\": container with ID starting with dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.952058 4809 scope.go:117] "RemoveContainer" containerID="d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.952290 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b"} err="failed to get container status \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": rpc error: code = NotFound desc = could not find container \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": container with ID starting with d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.952313 4809 scope.go:117] "RemoveContainer" containerID="2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.952532 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d"} err="failed to get container status \"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\": rpc error: code = NotFound desc = could not find container \"2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d\": container with ID starting with 2f89ac12060b8ca51279715ed67ea9d8feb8c29f671ee0e715def5f22e693c3d not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.952557 4809 scope.go:117] "RemoveContainer" containerID="0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.952794 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8"} err="failed to get container status \"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\": rpc error: code = NotFound desc = could not find container \"0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8\": container with ID starting with 0db0568a6a806743da70b28603e007af44a21742226fd272b2b7412365dd0ba8 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.952816 4809 scope.go:117] "RemoveContainer" containerID="d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.953030 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1"} err="failed to get container status \"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\": rpc error: code = NotFound desc = could not find container \"d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1\": container with ID starting with d0b292a57214462292e9c83ffed6cfa178c1dea95eea0629947fd2650e4727a1 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.953051 4809 scope.go:117] "RemoveContainer" containerID="31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.953255 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45"} err="failed to get container status \"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\": rpc error: code = NotFound desc = could not find container \"31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45\": container with ID starting with 31622777b2b08ead30db1178551b57659b665324e7297e2fde392f518fb9ec45 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.953277 4809 scope.go:117] "RemoveContainer" containerID="7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.953486 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9"} err="failed to get container status \"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\": rpc error: code = NotFound desc = could not find container \"7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9\": container with ID starting with 7d132094278d7ef17bab401a51f35a688c2d2bf75570118a1ff76ae033dce0f9 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.953510 4809 scope.go:117] "RemoveContainer" containerID="ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.953738 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc"} err="failed to get container status \"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\": rpc error: code = NotFound desc = could not find container \"ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc\": container with ID starting with ec8afe9204fa1590fc10bd65b07e3dff193c2b58447483c03137bc3385b701bc not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.953789 4809 scope.go:117] "RemoveContainer" containerID="b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.954015 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3"} err="failed to get container status \"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\": rpc error: code = NotFound desc = could not find container \"b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3\": container with ID starting with b574322ce140e0adfeebfc4d990d2e26a172b864d78bf7e62b2648c1c35471e3 not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.954038 4809 scope.go:117] "RemoveContainer" containerID="a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.954302 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a"} err="failed to get container status \"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\": rpc error: code = NotFound desc = could not find container \"a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a\": container with ID starting with a99294fc0c75cd867537064df057d85b150f52a14c495eaf94ad821df6c4c79a not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.954331 4809 scope.go:117] "RemoveContainer" containerID="dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.954584 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f"} err="failed to get container status \"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\": rpc error: code = NotFound desc = could not find container \"dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f\": container with ID starting with dfeec32cd309f577d2bbcdf8d4620b887bca5925615394167e919a072c55537f not found: ID does not exist" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.954604 4809 scope.go:117] "RemoveContainer" containerID="d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b" Nov 27 17:21:51 crc kubenswrapper[4809]: I1127 17:21:51.954813 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b"} err="failed to get container status \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": rpc error: code = NotFound desc = could not find container \"d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b\": container with ID starting with d2cd13feb152b9a565782a33eed0d6c5ff7ed2f0aab263d65e932bb8ed6ff22b not found: ID does not exist" Nov 27 17:21:52 crc kubenswrapper[4809]: I1127 17:21:52.672722 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerStarted","Data":"0a4d7374a7c9b7af7f9460b098b70705582def707e7fc0b92545d8ba324af2f2"} Nov 27 17:21:52 crc kubenswrapper[4809]: I1127 17:21:52.673105 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerStarted","Data":"12d962555bb0b32e78fb047a1816aab7a6b217f3d036179286e3cf65824ff5cc"} Nov 27 17:21:52 crc kubenswrapper[4809]: I1127 17:21:52.673116 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerStarted","Data":"959e1badd09ca5191507e076880f64aca03c0ce218a3800ddb68005622881d6d"} Nov 27 17:21:52 crc kubenswrapper[4809]: I1127 17:21:52.673126 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerStarted","Data":"0834334bb2a3a9238844dcd2e3bb6688a9270ca45019d8ec87610c60316d68a7"} Nov 27 17:21:52 crc kubenswrapper[4809]: I1127 17:21:52.673137 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerStarted","Data":"4d6385c08ebe9928f996e16809401864f80a31d46f7fda475cec357d4377f03a"} Nov 27 17:21:52 crc kubenswrapper[4809]: I1127 17:21:52.673146 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerStarted","Data":"00631c999e834ccfa92d59a4d556b749cef8f4dac092a7d369aedee82279e6aa"} Nov 27 17:21:53 crc kubenswrapper[4809]: I1127 17:21:53.467414 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b4318a-f089-451d-8a16-97de26acce28" path="/var/lib/kubelet/pods/66b4318a-f089-451d-8a16-97de26acce28/volumes" Nov 27 17:21:54 crc kubenswrapper[4809]: I1127 17:21:54.688655 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerStarted","Data":"9f5eae7de018c3972ff7896fa949c0c966530aa53d6a636bc4d4cfb3e03293cd"} Nov 27 17:21:55 crc kubenswrapper[4809]: I1127 17:21:55.409066 4809 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 17:21:57 crc kubenswrapper[4809]: I1127 17:21:57.710824 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" event={"ID":"49b29b48-c4c5-403b-ba79-21c18dda8430","Type":"ContainerStarted","Data":"817c2092e1012896fb0aa0f90e91b7d747c106c00386ebbb23a97986c0d2afe3"} Nov 27 17:21:57 crc kubenswrapper[4809]: I1127 17:21:57.711865 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:57 crc kubenswrapper[4809]: I1127 17:21:57.711895 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:57 crc kubenswrapper[4809]: I1127 17:21:57.711915 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:57 crc kubenswrapper[4809]: I1127 17:21:57.758351 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" podStartSLOduration=7.758327251 podStartE2EDuration="7.758327251s" podCreationTimestamp="2025-11-27 17:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:21:57.754381623 +0000 UTC m=+753.026838975" watchObservedRunningTime="2025-11-27 17:21:57.758327251 +0000 UTC m=+753.030784593" Nov 27 17:21:57 crc kubenswrapper[4809]: I1127 17:21:57.766533 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:21:57 crc kubenswrapper[4809]: I1127 17:21:57.784306 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.719431 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr"] Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.721062 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.724329 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.730676 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr"] Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.877663 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.877728 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.877795 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz79c\" (UniqueName: \"kubernetes.io/projected/57adfa91-4c88-4dc8-a92c-110ef12b330c-kube-api-access-wz79c\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.904365 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf"] Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.905760 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.911481 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf"] Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.979473 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.979563 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.979620 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz79c\" (UniqueName: \"kubernetes.io/projected/57adfa91-4c88-4dc8-a92c-110ef12b330c-kube-api-access-wz79c\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.980135 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.980217 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:13 crc kubenswrapper[4809]: I1127 17:22:13.998172 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz79c\" (UniqueName: \"kubernetes.io/projected/57adfa91-4c88-4dc8-a92c-110ef12b330c-kube-api-access-wz79c\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.047128 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.081437 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjwks\" (UniqueName: \"kubernetes.io/projected/88f51563-f860-48b1-b0e5-f5c11bfc78e5-kube-api-access-qjwks\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.081524 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.081550 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.182410 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.182929 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.183040 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjwks\" (UniqueName: \"kubernetes.io/projected/88f51563-f860-48b1-b0e5-f5c11bfc78e5-kube-api-access-qjwks\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.183191 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.183437 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.218051 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjwks\" (UniqueName: \"kubernetes.io/projected/88f51563-f860-48b1-b0e5-f5c11bfc78e5-kube-api-access-qjwks\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.228246 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.296294 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr"] Nov 27 17:22:14 crc kubenswrapper[4809]: W1127 17:22:14.300664 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57adfa91_4c88_4dc8_a92c_110ef12b330c.slice/crio-6c6a4e4b70ccb5449abc69f116633746cdc0919a4b6f0275bc7536306814cf92 WatchSource:0}: Error finding container 6c6a4e4b70ccb5449abc69f116633746cdc0919a4b6f0275bc7536306814cf92: Status 404 returned error can't find the container with id 6c6a4e4b70ccb5449abc69f116633746cdc0919a4b6f0275bc7536306814cf92 Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.647238 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf"] Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.827694 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" event={"ID":"88f51563-f860-48b1-b0e5-f5c11bfc78e5","Type":"ContainerStarted","Data":"b957ba48e1d10eb1baf4bcb8932974a118e58749fdbe3b2b1ff77c5258f10046"} Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.830358 4809 generic.go:334] "Generic (PLEG): container finished" podID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerID="12b8d7f66e8ede38a941826bb3e002ea2c0b2b2b8f3783bd3ae6fd25cbe87795" exitCode=0 Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.830440 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" event={"ID":"57adfa91-4c88-4dc8-a92c-110ef12b330c","Type":"ContainerDied","Data":"12b8d7f66e8ede38a941826bb3e002ea2c0b2b2b8f3783bd3ae6fd25cbe87795"} Nov 27 17:22:14 crc kubenswrapper[4809]: I1127 17:22:14.830483 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" event={"ID":"57adfa91-4c88-4dc8-a92c-110ef12b330c","Type":"ContainerStarted","Data":"6c6a4e4b70ccb5449abc69f116633746cdc0919a4b6f0275bc7536306814cf92"} Nov 27 17:22:14 crc kubenswrapper[4809]: E1127 17:22:14.838427 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57adfa91_4c88_4dc8_a92c_110ef12b330c.slice/crio-12b8d7f66e8ede38a941826bb3e002ea2c0b2b2b8f3783bd3ae6fd25cbe87795.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57adfa91_4c88_4dc8_a92c_110ef12b330c.slice/crio-conmon-12b8d7f66e8ede38a941826bb3e002ea2c0b2b2b8f3783bd3ae6fd25cbe87795.scope\": RecentStats: unable to find data in memory cache]" Nov 27 17:22:15 crc kubenswrapper[4809]: I1127 17:22:15.838213 4809 generic.go:334] "Generic (PLEG): container finished" podID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerID="63971d2b1732f89c36af664e2ee23b5b2331e4d2d559ae61c3cd6180d32ab6b4" exitCode=0 Nov 27 17:22:15 crc kubenswrapper[4809]: I1127 17:22:15.838320 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" event={"ID":"88f51563-f860-48b1-b0e5-f5c11bfc78e5","Type":"ContainerDied","Data":"63971d2b1732f89c36af664e2ee23b5b2331e4d2d559ae61c3cd6180d32ab6b4"} Nov 27 17:22:16 crc kubenswrapper[4809]: I1127 17:22:16.848977 4809 generic.go:334] "Generic (PLEG): container finished" podID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerID="c685e71e36534e3ee49bd908df32f0a7be883fd0809d55cee320cfecee1ae6fe" exitCode=0 Nov 27 17:22:16 crc kubenswrapper[4809]: I1127 17:22:16.849055 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" event={"ID":"57adfa91-4c88-4dc8-a92c-110ef12b330c","Type":"ContainerDied","Data":"c685e71e36534e3ee49bd908df32f0a7be883fd0809d55cee320cfecee1ae6fe"} Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.466499 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nv75s"] Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.468208 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.482517 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nv75s"] Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.628816 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-utilities\") pod \"redhat-operators-nv75s\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.628908 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-catalog-content\") pod \"redhat-operators-nv75s\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.628996 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8m4w\" (UniqueName: \"kubernetes.io/projected/149fc446-19a5-4a7b-ae5d-23493e518782-kube-api-access-l8m4w\") pod \"redhat-operators-nv75s\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.730258 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-catalog-content\") pod \"redhat-operators-nv75s\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.730354 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8m4w\" (UniqueName: \"kubernetes.io/projected/149fc446-19a5-4a7b-ae5d-23493e518782-kube-api-access-l8m4w\") pod \"redhat-operators-nv75s\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.730403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-utilities\") pod \"redhat-operators-nv75s\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.730972 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-utilities\") pod \"redhat-operators-nv75s\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.731119 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-catalog-content\") pod \"redhat-operators-nv75s\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.749237 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8m4w\" (UniqueName: \"kubernetes.io/projected/149fc446-19a5-4a7b-ae5d-23493e518782-kube-api-access-l8m4w\") pod \"redhat-operators-nv75s\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.794337 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.860998 4809 generic.go:334] "Generic (PLEG): container finished" podID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerID="5be04991360112cd9f3739ab2219da8058377222e3ecfe5dc987797c1495cb5c" exitCode=0 Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.861098 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" event={"ID":"57adfa91-4c88-4dc8-a92c-110ef12b330c","Type":"ContainerDied","Data":"5be04991360112cd9f3739ab2219da8058377222e3ecfe5dc987797c1495cb5c"} Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.864524 4809 generic.go:334] "Generic (PLEG): container finished" podID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerID="75614c86950abf0ee4617396011e85116532be6765626db5c2d9e220b2cf5681" exitCode=0 Nov 27 17:22:17 crc kubenswrapper[4809]: I1127 17:22:17.864668 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" event={"ID":"88f51563-f860-48b1-b0e5-f5c11bfc78e5","Type":"ContainerDied","Data":"75614c86950abf0ee4617396011e85116532be6765626db5c2d9e220b2cf5681"} Nov 27 17:22:18 crc kubenswrapper[4809]: I1127 17:22:18.021309 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nv75s"] Nov 27 17:22:18 crc kubenswrapper[4809]: W1127 17:22:18.029876 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod149fc446_19a5_4a7b_ae5d_23493e518782.slice/crio-a3ab9a87030f5b19c37963a23f289111a4fdbe62b60f805ded295cafb293fe53 WatchSource:0}: Error finding container a3ab9a87030f5b19c37963a23f289111a4fdbe62b60f805ded295cafb293fe53: Status 404 returned error can't find the container with id a3ab9a87030f5b19c37963a23f289111a4fdbe62b60f805ded295cafb293fe53 Nov 27 17:22:18 crc kubenswrapper[4809]: I1127 17:22:18.881528 4809 generic.go:334] "Generic (PLEG): container finished" podID="149fc446-19a5-4a7b-ae5d-23493e518782" containerID="a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67" exitCode=0 Nov 27 17:22:18 crc kubenswrapper[4809]: I1127 17:22:18.881638 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv75s" event={"ID":"149fc446-19a5-4a7b-ae5d-23493e518782","Type":"ContainerDied","Data":"a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67"} Nov 27 17:22:18 crc kubenswrapper[4809]: I1127 17:22:18.881680 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv75s" event={"ID":"149fc446-19a5-4a7b-ae5d-23493e518782","Type":"ContainerStarted","Data":"a3ab9a87030f5b19c37963a23f289111a4fdbe62b60f805ded295cafb293fe53"} Nov 27 17:22:18 crc kubenswrapper[4809]: I1127 17:22:18.885116 4809 generic.go:334] "Generic (PLEG): container finished" podID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerID="9bc85c58f44901adac8e32062f4798528a77e3e5c866710514bb560e4922f5d4" exitCode=0 Nov 27 17:22:18 crc kubenswrapper[4809]: I1127 17:22:18.885182 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" event={"ID":"88f51563-f860-48b1-b0e5-f5c11bfc78e5","Type":"ContainerDied","Data":"9bc85c58f44901adac8e32062f4798528a77e3e5c866710514bb560e4922f5d4"} Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.104931 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.248284 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-bundle\") pod \"57adfa91-4c88-4dc8-a92c-110ef12b330c\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.248410 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz79c\" (UniqueName: \"kubernetes.io/projected/57adfa91-4c88-4dc8-a92c-110ef12b330c-kube-api-access-wz79c\") pod \"57adfa91-4c88-4dc8-a92c-110ef12b330c\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.248466 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-util\") pod \"57adfa91-4c88-4dc8-a92c-110ef12b330c\" (UID: \"57adfa91-4c88-4dc8-a92c-110ef12b330c\") " Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.251733 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-bundle" (OuterVolumeSpecName: "bundle") pod "57adfa91-4c88-4dc8-a92c-110ef12b330c" (UID: "57adfa91-4c88-4dc8-a92c-110ef12b330c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.254945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57adfa91-4c88-4dc8-a92c-110ef12b330c-kube-api-access-wz79c" (OuterVolumeSpecName: "kube-api-access-wz79c") pod "57adfa91-4c88-4dc8-a92c-110ef12b330c" (UID: "57adfa91-4c88-4dc8-a92c-110ef12b330c"). InnerVolumeSpecName "kube-api-access-wz79c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.262465 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-util" (OuterVolumeSpecName: "util") pod "57adfa91-4c88-4dc8-a92c-110ef12b330c" (UID: "57adfa91-4c88-4dc8-a92c-110ef12b330c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.350346 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.350393 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz79c\" (UniqueName: \"kubernetes.io/projected/57adfa91-4c88-4dc8-a92c-110ef12b330c-kube-api-access-wz79c\") on node \"crc\" DevicePath \"\"" Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.350404 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57adfa91-4c88-4dc8-a92c-110ef12b330c-util\") on node \"crc\" DevicePath \"\"" Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.893048 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" event={"ID":"57adfa91-4c88-4dc8-a92c-110ef12b330c","Type":"ContainerDied","Data":"6c6a4e4b70ccb5449abc69f116633746cdc0919a4b6f0275bc7536306814cf92"} Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.893104 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c6a4e4b70ccb5449abc69f116633746cdc0919a4b6f0275bc7536306814cf92" Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.893199 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr" Nov 27 17:22:19 crc kubenswrapper[4809]: I1127 17:22:19.896654 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv75s" event={"ID":"149fc446-19a5-4a7b-ae5d-23493e518782","Type":"ContainerStarted","Data":"9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09"} Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.155923 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.262345 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-util\") pod \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.262418 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjwks\" (UniqueName: \"kubernetes.io/projected/88f51563-f860-48b1-b0e5-f5c11bfc78e5-kube-api-access-qjwks\") pod \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.262447 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-bundle\") pod \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\" (UID: \"88f51563-f860-48b1-b0e5-f5c11bfc78e5\") " Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.263435 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-bundle" (OuterVolumeSpecName: "bundle") pod "88f51563-f860-48b1-b0e5-f5c11bfc78e5" (UID: "88f51563-f860-48b1-b0e5-f5c11bfc78e5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.266409 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88f51563-f860-48b1-b0e5-f5c11bfc78e5-kube-api-access-qjwks" (OuterVolumeSpecName: "kube-api-access-qjwks") pod "88f51563-f860-48b1-b0e5-f5c11bfc78e5" (UID: "88f51563-f860-48b1-b0e5-f5c11bfc78e5"). InnerVolumeSpecName "kube-api-access-qjwks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.353946 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-util" (OuterVolumeSpecName: "util") pod "88f51563-f860-48b1-b0e5-f5c11bfc78e5" (UID: "88f51563-f860-48b1-b0e5-f5c11bfc78e5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.364642 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-util\") on node \"crc\" DevicePath \"\"" Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.364704 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjwks\" (UniqueName: \"kubernetes.io/projected/88f51563-f860-48b1-b0e5-f5c11bfc78e5-kube-api-access-qjwks\") on node \"crc\" DevicePath \"\"" Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.364722 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/88f51563-f860-48b1-b0e5-f5c11bfc78e5-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.902626 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" event={"ID":"88f51563-f860-48b1-b0e5-f5c11bfc78e5","Type":"ContainerDied","Data":"b957ba48e1d10eb1baf4bcb8932974a118e58749fdbe3b2b1ff77c5258f10046"} Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.902668 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b957ba48e1d10eb1baf4bcb8932974a118e58749fdbe3b2b1ff77c5258f10046" Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.902670 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf" Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.905162 4809 generic.go:334] "Generic (PLEG): container finished" podID="149fc446-19a5-4a7b-ae5d-23493e518782" containerID="9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09" exitCode=0 Nov 27 17:22:20 crc kubenswrapper[4809]: I1127 17:22:20.905188 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv75s" event={"ID":"149fc446-19a5-4a7b-ae5d-23493e518782","Type":"ContainerDied","Data":"9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09"} Nov 27 17:22:21 crc kubenswrapper[4809]: I1127 17:22:21.271043 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xtll8" Nov 27 17:22:21 crc kubenswrapper[4809]: I1127 17:22:21.913615 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv75s" event={"ID":"149fc446-19a5-4a7b-ae5d-23493e518782","Type":"ContainerStarted","Data":"acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c"} Nov 27 17:22:21 crc kubenswrapper[4809]: I1127 17:22:21.940270 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nv75s" podStartSLOduration=2.532071726 podStartE2EDuration="4.940244955s" podCreationTimestamp="2025-11-27 17:22:17 +0000 UTC" firstStartedPulling="2025-11-27 17:22:18.884140106 +0000 UTC m=+774.156597458" lastFinishedPulling="2025-11-27 17:22:21.292313335 +0000 UTC m=+776.564770687" observedRunningTime="2025-11-27 17:22:21.936012779 +0000 UTC m=+777.208470131" watchObservedRunningTime="2025-11-27 17:22:21.940244955 +0000 UTC m=+777.212702317" Nov 27 17:22:25 crc kubenswrapper[4809]: I1127 17:22:25.779184 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:22:25 crc kubenswrapper[4809]: I1127 17:22:25.779536 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.761723 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7"] Nov 27 17:22:27 crc kubenswrapper[4809]: E1127 17:22:27.762279 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerName="extract" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.762295 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerName="extract" Nov 27 17:22:27 crc kubenswrapper[4809]: E1127 17:22:27.762315 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerName="util" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.762322 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerName="util" Nov 27 17:22:27 crc kubenswrapper[4809]: E1127 17:22:27.762337 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerName="pull" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.762345 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerName="pull" Nov 27 17:22:27 crc kubenswrapper[4809]: E1127 17:22:27.762353 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerName="pull" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.762359 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerName="pull" Nov 27 17:22:27 crc kubenswrapper[4809]: E1127 17:22:27.762368 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerName="util" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.762374 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerName="util" Nov 27 17:22:27 crc kubenswrapper[4809]: E1127 17:22:27.762380 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerName="extract" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.762386 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerName="extract" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.762483 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="57adfa91-4c88-4dc8-a92c-110ef12b330c" containerName="extract" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.762493 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" containerName="extract" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.763224 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.765618 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-fwjrn" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.765873 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.765996 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.766351 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.768195 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.780719 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.792936 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7"] Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.794520 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.794840 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.848561 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.861602 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwrlp\" (UniqueName: \"kubernetes.io/projected/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-kube-api-access-nwrlp\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.861686 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.861718 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-apiservice-cert\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.861771 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-webhook-cert\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.861821 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-manager-config\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.963157 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-apiservice-cert\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.963497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-webhook-cert\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.963648 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-manager-config\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.963810 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwrlp\" (UniqueName: \"kubernetes.io/projected/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-kube-api-access-nwrlp\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.963934 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.964752 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-manager-config\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.971907 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-webhook-cert\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.972538 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.973128 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-apiservice-cert\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:27 crc kubenswrapper[4809]: I1127 17:22:27.997216 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwrlp\" (UniqueName: \"kubernetes.io/projected/2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61-kube-api-access-nwrlp\") pod \"loki-operator-controller-manager-fbb7c7d9f-bnjq7\" (UID: \"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61\") " pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:28 crc kubenswrapper[4809]: I1127 17:22:28.014666 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:28 crc kubenswrapper[4809]: I1127 17:22:28.084274 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:28 crc kubenswrapper[4809]: I1127 17:22:28.413987 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7"] Nov 27 17:22:28 crc kubenswrapper[4809]: W1127 17:22:28.421561 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f358eb9_3e6f_4efb_bd8f_6ef3320f3a61.slice/crio-7c68d882081c24c0557256b1fa7e611f8db744ad08076521ad47c32b67b21088 WatchSource:0}: Error finding container 7c68d882081c24c0557256b1fa7e611f8db744ad08076521ad47c32b67b21088: Status 404 returned error can't find the container with id 7c68d882081c24c0557256b1fa7e611f8db744ad08076521ad47c32b67b21088 Nov 27 17:22:28 crc kubenswrapper[4809]: I1127 17:22:28.854105 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nv75s"] Nov 27 17:22:28 crc kubenswrapper[4809]: I1127 17:22:28.958189 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" event={"ID":"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61","Type":"ContainerStarted","Data":"7c68d882081c24c0557256b1fa7e611f8db744ad08076521ad47c32b67b21088"} Nov 27 17:22:29 crc kubenswrapper[4809]: I1127 17:22:29.972931 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nv75s" podUID="149fc446-19a5-4a7b-ae5d-23493e518782" containerName="registry-server" containerID="cri-o://acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c" gracePeriod=2 Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.694665 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.808470 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8m4w\" (UniqueName: \"kubernetes.io/projected/149fc446-19a5-4a7b-ae5d-23493e518782-kube-api-access-l8m4w\") pod \"149fc446-19a5-4a7b-ae5d-23493e518782\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.808634 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-catalog-content\") pod \"149fc446-19a5-4a7b-ae5d-23493e518782\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.808733 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-utilities\") pod \"149fc446-19a5-4a7b-ae5d-23493e518782\" (UID: \"149fc446-19a5-4a7b-ae5d-23493e518782\") " Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.809832 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-utilities" (OuterVolumeSpecName: "utilities") pod "149fc446-19a5-4a7b-ae5d-23493e518782" (UID: "149fc446-19a5-4a7b-ae5d-23493e518782"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.828010 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/149fc446-19a5-4a7b-ae5d-23493e518782-kube-api-access-l8m4w" (OuterVolumeSpecName: "kube-api-access-l8m4w") pod "149fc446-19a5-4a7b-ae5d-23493e518782" (UID: "149fc446-19a5-4a7b-ae5d-23493e518782"). InnerVolumeSpecName "kube-api-access-l8m4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.911122 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8m4w\" (UniqueName: \"kubernetes.io/projected/149fc446-19a5-4a7b-ae5d-23493e518782-kube-api-access-l8m4w\") on node \"crc\" DevicePath \"\"" Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.911178 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.933710 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "149fc446-19a5-4a7b-ae5d-23493e518782" (UID: "149fc446-19a5-4a7b-ae5d-23493e518782"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.990004 4809 generic.go:334] "Generic (PLEG): container finished" podID="149fc446-19a5-4a7b-ae5d-23493e518782" containerID="acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c" exitCode=0 Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.990072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv75s" event={"ID":"149fc446-19a5-4a7b-ae5d-23493e518782","Type":"ContainerDied","Data":"acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c"} Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.990109 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv75s" event={"ID":"149fc446-19a5-4a7b-ae5d-23493e518782","Type":"ContainerDied","Data":"a3ab9a87030f5b19c37963a23f289111a4fdbe62b60f805ded295cafb293fe53"} Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.990135 4809 scope.go:117] "RemoveContainer" containerID="acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c" Nov 27 17:22:30 crc kubenswrapper[4809]: I1127 17:22:30.990339 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv75s" Nov 27 17:22:31 crc kubenswrapper[4809]: I1127 17:22:31.012061 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149fc446-19a5-4a7b-ae5d-23493e518782-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:22:31 crc kubenswrapper[4809]: I1127 17:22:31.019783 4809 scope.go:117] "RemoveContainer" containerID="9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09" Nov 27 17:22:31 crc kubenswrapper[4809]: I1127 17:22:31.031624 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nv75s"] Nov 27 17:22:31 crc kubenswrapper[4809]: I1127 17:22:31.034907 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nv75s"] Nov 27 17:22:31 crc kubenswrapper[4809]: I1127 17:22:31.467198 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="149fc446-19a5-4a7b-ae5d-23493e518782" path="/var/lib/kubelet/pods/149fc446-19a5-4a7b-ae5d-23493e518782/volumes" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.169832 4809 scope.go:117] "RemoveContainer" containerID="a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.216943 4809 scope.go:117] "RemoveContainer" containerID="acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c" Nov 27 17:22:33 crc kubenswrapper[4809]: E1127 17:22:33.217630 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c\": container with ID starting with acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c not found: ID does not exist" containerID="acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.217699 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c"} err="failed to get container status \"acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c\": rpc error: code = NotFound desc = could not find container \"acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c\": container with ID starting with acbc80c0cbaa607404bde98cddd6dccbaaa8dfc35f1e67c783887ffc41d4207c not found: ID does not exist" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.217771 4809 scope.go:117] "RemoveContainer" containerID="9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09" Nov 27 17:22:33 crc kubenswrapper[4809]: E1127 17:22:33.218200 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09\": container with ID starting with 9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09 not found: ID does not exist" containerID="9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.218260 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09"} err="failed to get container status \"9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09\": rpc error: code = NotFound desc = could not find container \"9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09\": container with ID starting with 9a9955623e2be8fdebdf4c210eb06e935237d0b31b8bea9e76b5298f4e2e1e09 not found: ID does not exist" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.218298 4809 scope.go:117] "RemoveContainer" containerID="a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67" Nov 27 17:22:33 crc kubenswrapper[4809]: E1127 17:22:33.219788 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67\": container with ID starting with a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67 not found: ID does not exist" containerID="a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.219864 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67"} err="failed to get container status \"a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67\": rpc error: code = NotFound desc = could not find container \"a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67\": container with ID starting with a0c67c8d338c3b41bcb1c07a47f297ff6fd9aa379cce102de2c788f8c827eb67 not found: ID does not exist" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.492889 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-8zwm2"] Nov 27 17:22:33 crc kubenswrapper[4809]: E1127 17:22:33.493882 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="149fc446-19a5-4a7b-ae5d-23493e518782" containerName="extract-utilities" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.493899 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="149fc446-19a5-4a7b-ae5d-23493e518782" containerName="extract-utilities" Nov 27 17:22:33 crc kubenswrapper[4809]: E1127 17:22:33.493919 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="149fc446-19a5-4a7b-ae5d-23493e518782" containerName="registry-server" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.493926 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="149fc446-19a5-4a7b-ae5d-23493e518782" containerName="registry-server" Nov 27 17:22:33 crc kubenswrapper[4809]: E1127 17:22:33.493948 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="149fc446-19a5-4a7b-ae5d-23493e518782" containerName="extract-content" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.493955 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="149fc446-19a5-4a7b-ae5d-23493e518782" containerName="extract-content" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.504491 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="149fc446-19a5-4a7b-ae5d-23493e518782" containerName="registry-server" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.507008 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-8zwm2" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.517532 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.517709 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.518278 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-j76w5" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.533385 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-8zwm2"] Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.649693 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k9l7\" (UniqueName: \"kubernetes.io/projected/c0697c42-98e3-49fd-b923-093806ec2876-kube-api-access-4k9l7\") pod \"cluster-logging-operator-ff9846bd-8zwm2\" (UID: \"c0697c42-98e3-49fd-b923-093806ec2876\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-8zwm2" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.751219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k9l7\" (UniqueName: \"kubernetes.io/projected/c0697c42-98e3-49fd-b923-093806ec2876-kube-api-access-4k9l7\") pod \"cluster-logging-operator-ff9846bd-8zwm2\" (UID: \"c0697c42-98e3-49fd-b923-093806ec2876\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-8zwm2" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.774326 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k9l7\" (UniqueName: \"kubernetes.io/projected/c0697c42-98e3-49fd-b923-093806ec2876-kube-api-access-4k9l7\") pod \"cluster-logging-operator-ff9846bd-8zwm2\" (UID: \"c0697c42-98e3-49fd-b923-093806ec2876\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-8zwm2" Nov 27 17:22:33 crc kubenswrapper[4809]: I1127 17:22:33.845107 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-8zwm2" Nov 27 17:22:34 crc kubenswrapper[4809]: I1127 17:22:34.021197 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" event={"ID":"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61","Type":"ContainerStarted","Data":"e6f1829366d3d5e8820a0039df57b0e41ae5497b0d5e5da8694377c62a8a18f8"} Nov 27 17:22:34 crc kubenswrapper[4809]: I1127 17:22:34.090985 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-8zwm2"] Nov 27 17:22:35 crc kubenswrapper[4809]: I1127 17:22:35.029828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-8zwm2" event={"ID":"c0697c42-98e3-49fd-b923-093806ec2876","Type":"ContainerStarted","Data":"1fe238356810c3781daea57133bb48539d5da32ed31d26dfed467328d93fee97"} Nov 27 17:22:42 crc kubenswrapper[4809]: I1127 17:22:42.119274 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-8zwm2" event={"ID":"c0697c42-98e3-49fd-b923-093806ec2876","Type":"ContainerStarted","Data":"b89dfdb31297b7cee04140ac0aef0da22b94db1b2f056d1dfb492c650a36dc31"} Nov 27 17:22:42 crc kubenswrapper[4809]: I1127 17:22:42.121897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" event={"ID":"2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61","Type":"ContainerStarted","Data":"77217bd06bb463f130256796c0c1676dbee30cbdf2be244a7b63aa622322a927"} Nov 27 17:22:42 crc kubenswrapper[4809]: I1127 17:22:42.122165 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:42 crc kubenswrapper[4809]: I1127 17:22:42.125802 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" Nov 27 17:22:42 crc kubenswrapper[4809]: I1127 17:22:42.139104 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-8zwm2" podStartSLOduration=1.7756251619999999 podStartE2EDuration="9.139084649s" podCreationTimestamp="2025-11-27 17:22:33 +0000 UTC" firstStartedPulling="2025-11-27 17:22:34.106889029 +0000 UTC m=+789.379346381" lastFinishedPulling="2025-11-27 17:22:41.470348516 +0000 UTC m=+796.742805868" observedRunningTime="2025-11-27 17:22:42.137789683 +0000 UTC m=+797.410247035" watchObservedRunningTime="2025-11-27 17:22:42.139084649 +0000 UTC m=+797.411542001" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.249702 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-fbb7c7d9f-bnjq7" podStartSLOduration=7.179614368 podStartE2EDuration="20.249679206s" podCreationTimestamp="2025-11-27 17:22:27 +0000 UTC" firstStartedPulling="2025-11-27 17:22:28.424146414 +0000 UTC m=+783.696603766" lastFinishedPulling="2025-11-27 17:22:41.494211252 +0000 UTC m=+796.766668604" observedRunningTime="2025-11-27 17:22:42.173999159 +0000 UTC m=+797.446456511" watchObservedRunningTime="2025-11-27 17:22:47.249679206 +0000 UTC m=+802.522136558" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.251118 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.251797 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.254340 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.254989 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.265205 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.365309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-59d39498-092e-4da4-bc3d-94392e35fcba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-59d39498-092e-4da4-bc3d-94392e35fcba\") pod \"minio\" (UID: \"120dd1ae-b077-48e5-8a74-c64d538c07af\") " pod="minio-dev/minio" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.365442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jzm2\" (UniqueName: \"kubernetes.io/projected/120dd1ae-b077-48e5-8a74-c64d538c07af-kube-api-access-6jzm2\") pod \"minio\" (UID: \"120dd1ae-b077-48e5-8a74-c64d538c07af\") " pod="minio-dev/minio" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.466600 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jzm2\" (UniqueName: \"kubernetes.io/projected/120dd1ae-b077-48e5-8a74-c64d538c07af-kube-api-access-6jzm2\") pod \"minio\" (UID: \"120dd1ae-b077-48e5-8a74-c64d538c07af\") " pod="minio-dev/minio" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.466691 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-59d39498-092e-4da4-bc3d-94392e35fcba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-59d39498-092e-4da4-bc3d-94392e35fcba\") pod \"minio\" (UID: \"120dd1ae-b077-48e5-8a74-c64d538c07af\") " pod="minio-dev/minio" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.471529 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.471583 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-59d39498-092e-4da4-bc3d-94392e35fcba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-59d39498-092e-4da4-bc3d-94392e35fcba\") pod \"minio\" (UID: \"120dd1ae-b077-48e5-8a74-c64d538c07af\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/55f50bb4fa1dc4165c177e0c5c9accc74b122ae773cb5237e5dcee2b15bdef39/globalmount\"" pod="minio-dev/minio" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.495594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jzm2\" (UniqueName: \"kubernetes.io/projected/120dd1ae-b077-48e5-8a74-c64d538c07af-kube-api-access-6jzm2\") pod \"minio\" (UID: \"120dd1ae-b077-48e5-8a74-c64d538c07af\") " pod="minio-dev/minio" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.498808 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-59d39498-092e-4da4-bc3d-94392e35fcba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-59d39498-092e-4da4-bc3d-94392e35fcba\") pod \"minio\" (UID: \"120dd1ae-b077-48e5-8a74-c64d538c07af\") " pod="minio-dev/minio" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.572055 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 27 17:22:47 crc kubenswrapper[4809]: I1127 17:22:47.796114 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 27 17:22:48 crc kubenswrapper[4809]: I1127 17:22:48.158017 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"120dd1ae-b077-48e5-8a74-c64d538c07af","Type":"ContainerStarted","Data":"ec291469a5f4784edcad39a12109c833d853b57eace9d632860a65cc8f7aaadc"} Nov 27 17:22:52 crc kubenswrapper[4809]: I1127 17:22:52.192379 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"120dd1ae-b077-48e5-8a74-c64d538c07af","Type":"ContainerStarted","Data":"935963d00ce287d4ff6c81956aee9ed79f6af97d504f8133bcd8724673310522"} Nov 27 17:22:52 crc kubenswrapper[4809]: I1127 17:22:52.211807 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.674621987 podStartE2EDuration="8.21178517s" podCreationTimestamp="2025-11-27 17:22:44 +0000 UTC" firstStartedPulling="2025-11-27 17:22:47.808344624 +0000 UTC m=+803.080801976" lastFinishedPulling="2025-11-27 17:22:51.345507807 +0000 UTC m=+806.617965159" observedRunningTime="2025-11-27 17:22:52.209446597 +0000 UTC m=+807.481903959" watchObservedRunningTime="2025-11-27 17:22:52.21178517 +0000 UTC m=+807.484242522" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.691769 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l"] Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.692713 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.695352 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.695549 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.695565 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-5fgnk" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.695662 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.696295 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.712387 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l"] Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.780221 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.780344 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.781618 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.781680 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.781730 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8dzs\" (UniqueName: \"kubernetes.io/projected/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-kube-api-access-x8dzs\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.781839 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-config\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.781882 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.843706 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-gr498"] Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.845124 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.850007 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.850114 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.850353 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.864740 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-gr498"] Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.883360 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.883463 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.883494 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.883532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8dzs\" (UniqueName: \"kubernetes.io/projected/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-kube-api-access-x8dzs\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.883564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-config\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.888320 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-config\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.888967 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.923926 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.925444 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.929364 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8dzs\" (UniqueName: \"kubernetes.io/projected/7d6e72ed-2edc-4744-8b78-adb9d95a41ee-kube-api-access-x8dzs\") pod \"logging-loki-distributor-76cc67bf56-wmb6l\" (UID: \"7d6e72ed-2edc-4744-8b78-adb9d95a41ee\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.974295 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq"] Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.975447 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.977991 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.978244 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.984817 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.984868 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.984917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.984944 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.984968 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79ms5\" (UniqueName: \"kubernetes.io/projected/ce25b13d-dcd3-420d-848a-afd09c1850f2-kube-api-access-79ms5\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.985071 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce25b13d-dcd3-420d-848a-afd09c1850f2-config\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:55 crc kubenswrapper[4809]: I1127 17:22:55.990930 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq"] Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.011464 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.069579 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-86596f6498-4d4pc"] Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.070681 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.073118 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.073409 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.073715 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.076846 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.078213 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086484 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce25b13d-dcd3-420d-848a-afd09c1850f2-config\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086541 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htcj5\" (UniqueName: \"kubernetes.io/projected/290cdb65-2424-4494-94cd-a69e5bbc126c-kube-api-access-htcj5\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086599 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/290cdb65-2424-4494-94cd-a69e5bbc126c-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086646 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086674 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086693 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/290cdb65-2424-4494-94cd-a69e5bbc126c-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086732 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79ms5\" (UniqueName: \"kubernetes.io/projected/ce25b13d-dcd3-420d-848a-afd09c1850f2-kube-api-access-79ms5\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086793 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/290cdb65-2424-4494-94cd-a69e5bbc126c-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.086819 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/290cdb65-2424-4494-94cd-a69e5bbc126c-config\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.087589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce25b13d-dcd3-420d-848a-afd09c1850f2-config\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.087653 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.088267 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-86596f6498-t5b62"] Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.091634 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.099609 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.103591 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-86596f6498-4d4pc"] Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.103711 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.106130 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-dz8xc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.107789 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-86596f6498-t5b62"] Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.110341 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/ce25b13d-dcd3-420d-848a-afd09c1850f2-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.115347 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79ms5\" (UniqueName: \"kubernetes.io/projected/ce25b13d-dcd3-420d-848a-afd09c1850f2-kube-api-access-79ms5\") pod \"logging-loki-querier-5895d59bb8-gr498\" (UID: \"ce25b13d-dcd3-420d-848a-afd09c1850f2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.178907 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.188556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-tenants\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.188786 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/290cdb65-2424-4494-94cd-a69e5bbc126c-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.188903 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/290cdb65-2424-4494-94cd-a69e5bbc126c-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.188933 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-logging-loki-ca-bundle\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.188952 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.188995 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-lokistack-gateway\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.189016 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-rbac\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.189032 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.189064 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdvsd\" (UniqueName: \"kubernetes.io/projected/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-kube-api-access-sdvsd\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.189716 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4239a407-a58d-46ac-a13b-0f744c3fbf0f-tenants\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.189849 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/290cdb65-2424-4494-94cd-a69e5bbc126c-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.189905 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-rbac\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.189946 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-tls-secret\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.189986 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmkvj\" (UniqueName: \"kubernetes.io/projected/4239a407-a58d-46ac-a13b-0f744c3fbf0f-kube-api-access-jmkvj\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.190042 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4239a407-a58d-46ac-a13b-0f744c3fbf0f-tls-secret\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.190133 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/290cdb65-2424-4494-94cd-a69e5bbc126c-config\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.190196 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-logging-loki-ca-bundle\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.190708 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.190787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-lokistack-gateway\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.190831 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4239a407-a58d-46ac-a13b-0f744c3fbf0f-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.190867 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htcj5\" (UniqueName: \"kubernetes.io/projected/290cdb65-2424-4494-94cd-a69e5bbc126c-kube-api-access-htcj5\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.191103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/290cdb65-2424-4494-94cd-a69e5bbc126c-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.191389 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/290cdb65-2424-4494-94cd-a69e5bbc126c-config\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.195649 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/290cdb65-2424-4494-94cd-a69e5bbc126c-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.195669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/290cdb65-2424-4494-94cd-a69e5bbc126c-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.209159 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htcj5\" (UniqueName: \"kubernetes.io/projected/290cdb65-2424-4494-94cd-a69e5bbc126c-kube-api-access-htcj5\") pod \"logging-loki-query-frontend-84558f7c9f-flnsq\" (UID: \"290cdb65-2424-4494-94cd-a69e5bbc126c\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.292898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-lokistack-gateway\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.292947 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-rbac\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.292975 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.292999 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdvsd\" (UniqueName: \"kubernetes.io/projected/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-kube-api-access-sdvsd\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293024 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4239a407-a58d-46ac-a13b-0f744c3fbf0f-tenants\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293047 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-rbac\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293064 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-tls-secret\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmkvj\" (UniqueName: \"kubernetes.io/projected/4239a407-a58d-46ac-a13b-0f744c3fbf0f-kube-api-access-jmkvj\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293108 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4239a407-a58d-46ac-a13b-0f744c3fbf0f-tls-secret\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293140 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-logging-loki-ca-bundle\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293165 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293195 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-lokistack-gateway\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293224 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4239a407-a58d-46ac-a13b-0f744c3fbf0f-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293256 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-tenants\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293300 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-logging-loki-ca-bundle\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.293319 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.294785 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-rbac\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.294805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.294953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-rbac\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.295253 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-lokistack-gateway\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.295525 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.295713 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.295731 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4239a407-a58d-46ac-a13b-0f744c3fbf0f-logging-loki-ca-bundle\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.297275 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4239a407-a58d-46ac-a13b-0f744c3fbf0f-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.298313 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4239a407-a58d-46ac-a13b-0f744c3fbf0f-tenants\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.298886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-lokistack-gateway\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.298886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-tenants\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.299671 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.300183 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4239a407-a58d-46ac-a13b-0f744c3fbf0f-tls-secret\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.300221 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-logging-loki-ca-bundle\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.306932 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-tls-secret\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.312393 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdvsd\" (UniqueName: \"kubernetes.io/projected/42bade25-cca5-4a1b-9c2f-67c1ffb97c68-kube-api-access-sdvsd\") pod \"logging-loki-gateway-86596f6498-t5b62\" (UID: \"42bade25-cca5-4a1b-9c2f-67c1ffb97c68\") " pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.314871 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmkvj\" (UniqueName: \"kubernetes.io/projected/4239a407-a58d-46ac-a13b-0f744c3fbf0f-kube-api-access-jmkvj\") pod \"logging-loki-gateway-86596f6498-4d4pc\" (UID: \"4239a407-a58d-46ac-a13b-0f744c3fbf0f\") " pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.415866 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.431684 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.507443 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l"] Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.513163 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq"] Nov 27 17:22:56 crc kubenswrapper[4809]: W1127 17:22:56.514608 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod290cdb65_2424_4494_94cd_a69e5bbc126c.slice/crio-663206d7499435a364d5c32f1ca165614ebc0a8efaf73434c432be883f144046 WatchSource:0}: Error finding container 663206d7499435a364d5c32f1ca165614ebc0a8efaf73434c432be883f144046: Status 404 returned error can't find the container with id 663206d7499435a364d5c32f1ca165614ebc0a8efaf73434c432be883f144046 Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.599361 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-gr498"] Nov 27 17:22:56 crc kubenswrapper[4809]: W1127 17:22:56.618113 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce25b13d_dcd3_420d_848a_afd09c1850f2.slice/crio-e6b2a2331a2e503c9398dc199b274526072212014cbf7891caa06ea968b0f9c3 WatchSource:0}: Error finding container e6b2a2331a2e503c9398dc199b274526072212014cbf7891caa06ea968b0f9c3: Status 404 returned error can't find the container with id e6b2a2331a2e503c9398dc199b274526072212014cbf7891caa06ea968b0f9c3 Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.734239 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-86596f6498-t5b62"] Nov 27 17:22:56 crc kubenswrapper[4809]: W1127 17:22:56.737300 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42bade25_cca5_4a1b_9c2f_67c1ffb97c68.slice/crio-84152631b23477c386e1c4dcadd46d352f7c93b9afa132f8731b235577745b5d WatchSource:0}: Error finding container 84152631b23477c386e1c4dcadd46d352f7c93b9afa132f8731b235577745b5d: Status 404 returned error can't find the container with id 84152631b23477c386e1c4dcadd46d352f7c93b9afa132f8731b235577745b5d Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.872847 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.873665 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.877935 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.878789 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.883609 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-86596f6498-4d4pc"] Nov 27 17:22:56 crc kubenswrapper[4809]: W1127 17:22:56.889709 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4239a407_a58d_46ac_a13b_0f744c3fbf0f.slice/crio-827505a7d5562a8730093b99cd65a51ea93d80cd60b81011129f50f934b8ead9 WatchSource:0}: Error finding container 827505a7d5562a8730093b99cd65a51ea93d80cd60b81011129f50f934b8ead9: Status 404 returned error can't find the container with id 827505a7d5562a8730093b99cd65a51ea93d80cd60b81011129f50f934b8ead9 Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.891911 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.930121 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.930908 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.933020 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.933405 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Nov 27 17:22:56 crc kubenswrapper[4809]: I1127 17:22:56.944564 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010019 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010085 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-196783dc-4256-411f-b531-3edbd05a98eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-196783dc-4256-411f-b531-3edbd05a98eb\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010122 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d8bf3ebc-a9bf-4f2a-aa35-e065588f2a42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d8bf3ebc-a9bf-4f2a-aa35-e065588f2a42\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010146 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsz29\" (UniqueName: \"kubernetes.io/projected/ed55809b-892d-4ba2-b037-5eb25f1add8d-kube-api-access-bsz29\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010178 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010197 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010216 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3f1cbe32-26f3-41dc-84cc-0bbe85f17385\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3f1cbe32-26f3-41dc-84cc-0bbe85f17385\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010754 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010863 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010941 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed55809b-892d-4ba2-b037-5eb25f1add8d-config\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.010990 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ecc35f-bd3c-43be-9841-0129299f73f9-config\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.011017 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.011047 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.011067 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbjqc\" (UniqueName: \"kubernetes.io/projected/87ecc35f-bd3c-43be-9841-0129299f73f9-kube-api-access-xbjqc\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.023288 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.024569 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.029036 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.029063 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.036015 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112454 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d8bf3ebc-a9bf-4f2a-aa35-e065588f2a42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d8bf3ebc-a9bf-4f2a-aa35-e065588f2a42\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112489 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsz29\" (UniqueName: \"kubernetes.io/projected/ed55809b-892d-4ba2-b037-5eb25f1add8d-kube-api-access-bsz29\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112526 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112576 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112601 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112625 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cccadb3e-654a-430f-9c69-c99a6fde2279-config\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112647 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112689 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3f1cbe32-26f3-41dc-84cc-0bbe85f17385\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3f1cbe32-26f3-41dc-84cc-0bbe85f17385\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112713 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xvjq\" (UniqueName: \"kubernetes.io/projected/cccadb3e-654a-430f-9c69-c99a6fde2279-kube-api-access-2xvjq\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112791 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112817 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-48691e39-df11-4000-afac-90af433f954a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48691e39-df11-4000-afac-90af433f954a\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112840 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed55809b-892d-4ba2-b037-5eb25f1add8d-config\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112910 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ecc35f-bd3c-43be-9841-0129299f73f9-config\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112935 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112965 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.112997 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbjqc\" (UniqueName: \"kubernetes.io/projected/87ecc35f-bd3c-43be-9841-0129299f73f9-kube-api-access-xbjqc\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.113026 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.113048 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-196783dc-4256-411f-b531-3edbd05a98eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-196783dc-4256-411f-b531-3edbd05a98eb\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.114430 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed55809b-892d-4ba2-b037-5eb25f1add8d-config\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.114660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ecc35f-bd3c-43be-9841-0129299f73f9-config\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.115046 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.115105 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.117611 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.117701 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-196783dc-4256-411f-b531-3edbd05a98eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-196783dc-4256-411f-b531-3edbd05a98eb\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/523610deb38e6af14c49ef8b04f124de05d26a9acf756001caa4ea01b50aa4c2/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.117788 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.118012 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.118098 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.118141 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3f1cbe32-26f3-41dc-84cc-0bbe85f17385\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3f1cbe32-26f3-41dc-84cc-0bbe85f17385\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/43535a4f3269bf9ae5fed4b76d34d564edab8b442fce0ca3f9d49726bfcde59e/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.118456 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.118721 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.118767 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d8bf3ebc-a9bf-4f2a-aa35-e065588f2a42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d8bf3ebc-a9bf-4f2a-aa35-e065588f2a42\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5354771e0cc130e4daec40e47b431b113f17f19e0f7cc72aba4226f69e38d2fa/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.120681 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/87ecc35f-bd3c-43be-9841-0129299f73f9-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.124944 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.125570 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/ed55809b-892d-4ba2-b037-5eb25f1add8d-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.135662 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsz29\" (UniqueName: \"kubernetes.io/projected/ed55809b-892d-4ba2-b037-5eb25f1add8d-kube-api-access-bsz29\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.135812 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbjqc\" (UniqueName: \"kubernetes.io/projected/87ecc35f-bd3c-43be-9841-0129299f73f9-kube-api-access-xbjqc\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.154230 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3f1cbe32-26f3-41dc-84cc-0bbe85f17385\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3f1cbe32-26f3-41dc-84cc-0bbe85f17385\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.155007 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-196783dc-4256-411f-b531-3edbd05a98eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-196783dc-4256-411f-b531-3edbd05a98eb\") pod \"logging-loki-ingester-0\" (UID: \"87ecc35f-bd3c-43be-9841-0129299f73f9\") " pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.164083 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d8bf3ebc-a9bf-4f2a-aa35-e065588f2a42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d8bf3ebc-a9bf-4f2a-aa35-e065588f2a42\") pod \"logging-loki-compactor-0\" (UID: \"ed55809b-892d-4ba2-b037-5eb25f1add8d\") " pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.194139 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.213847 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.213978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.214014 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.214033 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cccadb3e-654a-430f-9c69-c99a6fde2279-config\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.214057 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xvjq\" (UniqueName: \"kubernetes.io/projected/cccadb3e-654a-430f-9c69-c99a6fde2279-kube-api-access-2xvjq\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.214086 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.214114 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-48691e39-df11-4000-afac-90af433f954a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48691e39-df11-4000-afac-90af433f954a\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.215160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.215771 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cccadb3e-654a-430f-9c69-c99a6fde2279-config\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.217580 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.217622 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-48691e39-df11-4000-afac-90af433f954a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48691e39-df11-4000-afac-90af433f954a\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8cd325a062bcbafa46b59bf623dda640114119ce2a23855d6221dad60ba2513d/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.218093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.219229 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.219419 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/cccadb3e-654a-430f-9c69-c99a6fde2279-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.221760 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" event={"ID":"290cdb65-2424-4494-94cd-a69e5bbc126c","Type":"ContainerStarted","Data":"663206d7499435a364d5c32f1ca165614ebc0a8efaf73434c432be883f144046"} Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.223305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" event={"ID":"4239a407-a58d-46ac-a13b-0f744c3fbf0f","Type":"ContainerStarted","Data":"827505a7d5562a8730093b99cd65a51ea93d80cd60b81011129f50f934b8ead9"} Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.225363 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" event={"ID":"42bade25-cca5-4a1b-9c2f-67c1ffb97c68","Type":"ContainerStarted","Data":"84152631b23477c386e1c4dcadd46d352f7c93b9afa132f8731b235577745b5d"} Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.226842 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" event={"ID":"ce25b13d-dcd3-420d-848a-afd09c1850f2","Type":"ContainerStarted","Data":"e6b2a2331a2e503c9398dc199b274526072212014cbf7891caa06ea968b0f9c3"} Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.228237 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" event={"ID":"7d6e72ed-2edc-4744-8b78-adb9d95a41ee","Type":"ContainerStarted","Data":"fa0d8975a498e473c4ef8b72997bcdc830ce9df636881e87fdd9b8f7efc4225e"} Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.237389 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xvjq\" (UniqueName: \"kubernetes.io/projected/cccadb3e-654a-430f-9c69-c99a6fde2279-kube-api-access-2xvjq\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.245324 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-48691e39-df11-4000-afac-90af433f954a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48691e39-df11-4000-afac-90af433f954a\") pod \"logging-loki-index-gateway-0\" (UID: \"cccadb3e-654a-430f-9c69-c99a6fde2279\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.266186 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.343797 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.412672 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.680342 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 27 17:22:57 crc kubenswrapper[4809]: W1127 17:22:57.685606 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded55809b_892d_4ba2_b037_5eb25f1add8d.slice/crio-5c1c74ca4aeed95ed9e929887dd715dde70e882df54a47190bda39108e976d66 WatchSource:0}: Error finding container 5c1c74ca4aeed95ed9e929887dd715dde70e882df54a47190bda39108e976d66: Status 404 returned error can't find the container with id 5c1c74ca4aeed95ed9e929887dd715dde70e882df54a47190bda39108e976d66 Nov 27 17:22:57 crc kubenswrapper[4809]: I1127 17:22:57.770395 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 27 17:22:57 crc kubenswrapper[4809]: W1127 17:22:57.777286 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcccadb3e_654a_430f_9c69_c99a6fde2279.slice/crio-82cbfb2025880c8230ad8804b57f9a5d8af1e8690c051b4c3dee8f3058f1a459 WatchSource:0}: Error finding container 82cbfb2025880c8230ad8804b57f9a5d8af1e8690c051b4c3dee8f3058f1a459: Status 404 returned error can't find the container with id 82cbfb2025880c8230ad8804b57f9a5d8af1e8690c051b4c3dee8f3058f1a459 Nov 27 17:22:58 crc kubenswrapper[4809]: I1127 17:22:58.243061 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"cccadb3e-654a-430f-9c69-c99a6fde2279","Type":"ContainerStarted","Data":"82cbfb2025880c8230ad8804b57f9a5d8af1e8690c051b4c3dee8f3058f1a459"} Nov 27 17:22:58 crc kubenswrapper[4809]: I1127 17:22:58.244729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"ed55809b-892d-4ba2-b037-5eb25f1add8d","Type":"ContainerStarted","Data":"5c1c74ca4aeed95ed9e929887dd715dde70e882df54a47190bda39108e976d66"} Nov 27 17:22:58 crc kubenswrapper[4809]: I1127 17:22:58.248226 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"87ecc35f-bd3c-43be-9841-0129299f73f9","Type":"ContainerStarted","Data":"8293d29639728fb54c9d5a99f36133f9dc4f516a44ebe0d6e5e43383c5f374ad"} Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.262517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" event={"ID":"42bade25-cca5-4a1b-9c2f-67c1ffb97c68","Type":"ContainerStarted","Data":"ef30ca43af4187dbbeb8d99e02bae7cc3a411b72360ce6249bf0beea35092b7a"} Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.264084 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" event={"ID":"ce25b13d-dcd3-420d-848a-afd09c1850f2","Type":"ContainerStarted","Data":"9e66417b65c53310619769b6c14cd1b8caf001a492f6825aae023bae49d5726f"} Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.264238 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.265540 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" event={"ID":"7d6e72ed-2edc-4744-8b78-adb9d95a41ee","Type":"ContainerStarted","Data":"b3c3f3b68dd2ef45a0357d57ffbda315dc501fa520144eb0ab1163baf348fb86"} Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.265655 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.266829 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"ed55809b-892d-4ba2-b037-5eb25f1add8d","Type":"ContainerStarted","Data":"89ff87ff6304d18126562df83b61e39bb65632de048a33027cdfc435e0eaf3a0"} Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.266951 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.268396 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"87ecc35f-bd3c-43be-9841-0129299f73f9","Type":"ContainerStarted","Data":"d1d8abd75c811645efcb2fe1a25728c6fc73a15a6264cf195cfedfe824529b97"} Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.268499 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.269734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"cccadb3e-654a-430f-9c69-c99a6fde2279","Type":"ContainerStarted","Data":"7dfc9c1a4574ccf5e824e0ceece32e3a8f4b078ff4d1062f660ff6dda1688ef7"} Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.269869 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.271605 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" event={"ID":"290cdb65-2424-4494-94cd-a69e5bbc126c","Type":"ContainerStarted","Data":"0c26fea43d4ccf1aa0f1bce512937b724e662466030cc35deb0daa7f5819af3d"} Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.271718 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.273378 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" event={"ID":"4239a407-a58d-46ac-a13b-0f744c3fbf0f","Type":"ContainerStarted","Data":"1dcb845e69aeae77d0d8a81c95afcc348a87a894a4376fcca1fcf0b43342fa00"} Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.283818 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" podStartSLOduration=2.538014504 podStartE2EDuration="5.283800144s" podCreationTimestamp="2025-11-27 17:22:55 +0000 UTC" firstStartedPulling="2025-11-27 17:22:56.625558142 +0000 UTC m=+811.898015494" lastFinishedPulling="2025-11-27 17:22:59.371343782 +0000 UTC m=+814.643801134" observedRunningTime="2025-11-27 17:23:00.283538847 +0000 UTC m=+815.555996209" watchObservedRunningTime="2025-11-27 17:23:00.283800144 +0000 UTC m=+815.556257516" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.308377 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" podStartSLOduration=2.440631756 podStartE2EDuration="5.308358829s" podCreationTimestamp="2025-11-27 17:22:55 +0000 UTC" firstStartedPulling="2025-11-27 17:22:56.519378353 +0000 UTC m=+811.791835705" lastFinishedPulling="2025-11-27 17:22:59.387105436 +0000 UTC m=+814.659562778" observedRunningTime="2025-11-27 17:23:00.304887393 +0000 UTC m=+815.577344735" watchObservedRunningTime="2025-11-27 17:23:00.308358829 +0000 UTC m=+815.580816181" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.323349 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" podStartSLOduration=2.526156169 podStartE2EDuration="5.323326831s" podCreationTimestamp="2025-11-27 17:22:55 +0000 UTC" firstStartedPulling="2025-11-27 17:22:56.512539336 +0000 UTC m=+811.784996688" lastFinishedPulling="2025-11-27 17:22:59.309709998 +0000 UTC m=+814.582167350" observedRunningTime="2025-11-27 17:23:00.321910191 +0000 UTC m=+815.594367543" watchObservedRunningTime="2025-11-27 17:23:00.323326831 +0000 UTC m=+815.595784203" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.339134 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=2.7321170500000003 podStartE2EDuration="4.339109605s" podCreationTimestamp="2025-11-27 17:22:56 +0000 UTC" firstStartedPulling="2025-11-27 17:22:57.780176902 +0000 UTC m=+813.052634254" lastFinishedPulling="2025-11-27 17:22:59.387169457 +0000 UTC m=+814.659626809" observedRunningTime="2025-11-27 17:23:00.337300845 +0000 UTC m=+815.609758207" watchObservedRunningTime="2025-11-27 17:23:00.339109605 +0000 UTC m=+815.611566957" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.376983 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.680486829 podStartE2EDuration="5.376958555s" podCreationTimestamp="2025-11-27 17:22:55 +0000 UTC" firstStartedPulling="2025-11-27 17:22:57.690821215 +0000 UTC m=+812.963278567" lastFinishedPulling="2025-11-27 17:22:59.387292921 +0000 UTC m=+814.659750293" observedRunningTime="2025-11-27 17:23:00.368518943 +0000 UTC m=+815.640976295" watchObservedRunningTime="2025-11-27 17:23:00.376958555 +0000 UTC m=+815.649415907" Nov 27 17:23:00 crc kubenswrapper[4809]: I1127 17:23:00.393140 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=3.4273168800000002 podStartE2EDuration="5.393115679s" podCreationTimestamp="2025-11-27 17:22:55 +0000 UTC" firstStartedPulling="2025-11-27 17:22:57.422416187 +0000 UTC m=+812.694873539" lastFinishedPulling="2025-11-27 17:22:59.388214976 +0000 UTC m=+814.660672338" observedRunningTime="2025-11-27 17:23:00.386472026 +0000 UTC m=+815.658929378" watchObservedRunningTime="2025-11-27 17:23:00.393115679 +0000 UTC m=+815.665573031" Nov 27 17:23:02 crc kubenswrapper[4809]: I1127 17:23:02.290797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" event={"ID":"4239a407-a58d-46ac-a13b-0f744c3fbf0f","Type":"ContainerStarted","Data":"c4ca1b385389020c03b2788117df6b83755e7f25097fc04a2d829d9b0a52ad5e"} Nov 27 17:23:02 crc kubenswrapper[4809]: I1127 17:23:02.291882 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:23:02 crc kubenswrapper[4809]: I1127 17:23:02.293983 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:23:02 crc kubenswrapper[4809]: I1127 17:23:02.294027 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" event={"ID":"42bade25-cca5-4a1b-9c2f-67c1ffb97c68","Type":"ContainerStarted","Data":"d826a8f2b1dd84c6a8347ea3a415f64caec7b07e6b269b5fcac6cd0458cc03bf"} Nov 27 17:23:02 crc kubenswrapper[4809]: I1127 17:23:02.302114 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:23:02 crc kubenswrapper[4809]: I1127 17:23:02.306888 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:23:02 crc kubenswrapper[4809]: I1127 17:23:02.315298 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" podStartSLOduration=1.723145285 podStartE2EDuration="6.315280408s" podCreationTimestamp="2025-11-27 17:22:56 +0000 UTC" firstStartedPulling="2025-11-27 17:22:56.894273559 +0000 UTC m=+812.166730911" lastFinishedPulling="2025-11-27 17:23:01.486408682 +0000 UTC m=+816.758866034" observedRunningTime="2025-11-27 17:23:02.313159279 +0000 UTC m=+817.585616681" watchObservedRunningTime="2025-11-27 17:23:02.315280408 +0000 UTC m=+817.587737760" Nov 27 17:23:02 crc kubenswrapper[4809]: I1127 17:23:02.371049 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" podStartSLOduration=1.6197710920000001 podStartE2EDuration="6.37103066s" podCreationTimestamp="2025-11-27 17:22:56 +0000 UTC" firstStartedPulling="2025-11-27 17:22:56.739808953 +0000 UTC m=+812.012266305" lastFinishedPulling="2025-11-27 17:23:01.491068521 +0000 UTC m=+816.763525873" observedRunningTime="2025-11-27 17:23:02.370785294 +0000 UTC m=+817.643242646" watchObservedRunningTime="2025-11-27 17:23:02.37103066 +0000 UTC m=+817.643488022" Nov 27 17:23:03 crc kubenswrapper[4809]: I1127 17:23:03.298691 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:23:03 crc kubenswrapper[4809]: I1127 17:23:03.298765 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:23:03 crc kubenswrapper[4809]: I1127 17:23:03.307078 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-86596f6498-t5b62" Nov 27 17:23:03 crc kubenswrapper[4809]: I1127 17:23:03.311822 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-86596f6498-4d4pc" Nov 27 17:23:16 crc kubenswrapper[4809]: I1127 17:23:16.017165 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-wmb6l" Nov 27 17:23:16 crc kubenswrapper[4809]: I1127 17:23:16.183964 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-gr498" Nov 27 17:23:16 crc kubenswrapper[4809]: I1127 17:23:16.304378 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-flnsq" Nov 27 17:23:17 crc kubenswrapper[4809]: I1127 17:23:17.202846 4809 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 27 17:23:17 crc kubenswrapper[4809]: I1127 17:23:17.202937 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="87ecc35f-bd3c-43be-9841-0129299f73f9" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 27 17:23:17 crc kubenswrapper[4809]: I1127 17:23:17.274856 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Nov 27 17:23:17 crc kubenswrapper[4809]: I1127 17:23:17.352299 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Nov 27 17:23:25 crc kubenswrapper[4809]: I1127 17:23:25.779261 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:23:25 crc kubenswrapper[4809]: I1127 17:23:25.779974 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:23:25 crc kubenswrapper[4809]: I1127 17:23:25.780034 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:23:25 crc kubenswrapper[4809]: I1127 17:23:25.780802 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4c5bbc3124a8972c9912c532472d23e13e06555c5f6e029a3731193652f77944"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:23:25 crc kubenswrapper[4809]: I1127 17:23:25.780861 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://4c5bbc3124a8972c9912c532472d23e13e06555c5f6e029a3731193652f77944" gracePeriod=600 Nov 27 17:23:26 crc kubenswrapper[4809]: I1127 17:23:26.443162 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="4c5bbc3124a8972c9912c532472d23e13e06555c5f6e029a3731193652f77944" exitCode=0 Nov 27 17:23:26 crc kubenswrapper[4809]: I1127 17:23:26.443256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"4c5bbc3124a8972c9912c532472d23e13e06555c5f6e029a3731193652f77944"} Nov 27 17:23:26 crc kubenswrapper[4809]: I1127 17:23:26.443911 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"f6a9abdc5c8929f8e6588fb217c1a8bad8386fe8d57f874960d924240fd36c12"} Nov 27 17:23:26 crc kubenswrapper[4809]: I1127 17:23:26.443954 4809 scope.go:117] "RemoveContainer" containerID="cb5f4afccabcb96c09b1f91bcfcd46aff8f41c149de91ed16add7b6ee4b7e98c" Nov 27 17:23:27 crc kubenswrapper[4809]: I1127 17:23:27.197762 4809 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 27 17:23:27 crc kubenswrapper[4809]: I1127 17:23:27.197814 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="87ecc35f-bd3c-43be-9841-0129299f73f9" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 27 17:23:37 crc kubenswrapper[4809]: I1127 17:23:37.199791 4809 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 27 17:23:37 crc kubenswrapper[4809]: I1127 17:23:37.200428 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="87ecc35f-bd3c-43be-9841-0129299f73f9" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 27 17:23:47 crc kubenswrapper[4809]: I1127 17:23:47.198681 4809 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 27 17:23:47 crc kubenswrapper[4809]: I1127 17:23:47.201018 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="87ecc35f-bd3c-43be-9841-0129299f73f9" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 27 17:23:57 crc kubenswrapper[4809]: I1127 17:23:57.201577 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.821698 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-mzcrc"] Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.823180 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.828397 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-2kcsg" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.828409 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.828517 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.828642 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.828802 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.835254 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.853654 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-mzcrc"] Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.899592 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-mzcrc"] Nov 27 17:24:16 crc kubenswrapper[4809]: E1127 17:24:16.900381 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-w9v6b metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-w9v6b metrics sa-token tmp trusted-ca]: context canceled" pod="openshift-logging/collector-mzcrc" podUID="db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968533 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-metrics\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-syslog-receiver\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968648 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-token\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config-openshift-service-cacrt\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968689 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-entrypoint\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968717 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-tmp\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968765 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-sa-token\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968788 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9v6b\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-kube-api-access-w9v6b\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968819 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-datadir\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968841 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:16 crc kubenswrapper[4809]: I1127 17:24:16.968867 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-trusted-ca\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.069939 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-tmp\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.069992 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9v6b\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-kube-api-access-w9v6b\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.070009 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-sa-token\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.070063 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-datadir\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.070398 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-datadir\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.070085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.070564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-trusted-ca\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.070707 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-metrics\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.070876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-syslog-receiver\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.070945 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-token\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: E1127 17:24:17.070979 4809 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Nov 27 17:24:17 crc kubenswrapper[4809]: E1127 17:24:17.071044 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-syslog-receiver podName:db0e77a2-b4bf-46e3-aa53-b5e048f8bb95 nodeName:}" failed. No retries permitted until 2025-11-27 17:24:17.571025431 +0000 UTC m=+892.843482783 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-syslog-receiver") pod "collector-mzcrc" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95") : secret "collector-syslog-receiver" not found Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.070986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config-openshift-service-cacrt\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.071121 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-entrypoint\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.071683 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-trusted-ca\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.072093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-entrypoint\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.072215 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config-openshift-service-cacrt\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.072688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.077429 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-tmp\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.077887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-token\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.078525 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-metrics\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.088841 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9v6b\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-kube-api-access-w9v6b\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.099131 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-sa-token\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.579410 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-syslog-receiver\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.583249 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-syslog-receiver\") pod \"collector-mzcrc\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.781367 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.791118 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-mzcrc" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.884138 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-datadir\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.884262 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config-openshift-service-cacrt\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.884317 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-sa-token\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.884386 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-datadir" (OuterVolumeSpecName: "datadir") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.884877 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.885020 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-syslog-receiver\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.885077 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-token\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.885223 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-trusted-ca\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.885293 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-metrics\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.885356 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9v6b\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-kube-api-access-w9v6b\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.885420 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.885526 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-entrypoint\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.885613 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-tmp\") pod \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\" (UID: \"db0e77a2-b4bf-46e3-aa53-b5e048f8bb95\") " Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.886073 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.886536 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.886823 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config" (OuterVolumeSpecName: "config") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.886903 4809 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-entrypoint\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.887049 4809 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-datadir\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.887124 4809 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.887187 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.888375 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-token" (OuterVolumeSpecName: "collector-token") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.889074 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-metrics" (OuterVolumeSpecName: "metrics") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.889156 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.889174 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-kube-api-access-w9v6b" (OuterVolumeSpecName: "kube-api-access-w9v6b") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "kube-api-access-w9v6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.890018 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-tmp" (OuterVolumeSpecName: "tmp") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.890127 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-sa-token" (OuterVolumeSpecName: "sa-token") pod "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" (UID: "db0e77a2-b4bf-46e3-aa53-b5e048f8bb95"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.988131 4809 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.988174 4809 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-collector-token\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.988184 4809 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.988192 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9v6b\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-kube-api-access-w9v6b\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.988203 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.988212 4809 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-tmp\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:17 crc kubenswrapper[4809]: I1127 17:24:17.988220 4809 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.790665 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-mzcrc" Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.833026 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-mzcrc"] Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.836904 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-mzcrc"] Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.852879 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-ndnhl"] Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.853835 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-ndnhl" Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.856194 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.857776 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.857931 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-2kcsg" Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.858912 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.859517 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.866983 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 27 17:24:18 crc kubenswrapper[4809]: I1127 17:24:18.868569 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-ndnhl"] Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.003917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-config\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.003977 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-metrics\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.004002 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-tmp\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.004051 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-collector-syslog-receiver\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.004083 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-trusted-ca\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.004236 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzjfp\" (UniqueName: \"kubernetes.io/projected/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-kube-api-access-tzjfp\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.004282 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-datadir\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.004378 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-sa-token\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.004465 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-collector-token\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.004497 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-entrypoint\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.004601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-config-openshift-service-cacrt\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-config\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105587 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-metrics\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105613 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-tmp\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105654 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-collector-syslog-receiver\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105680 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-trusted-ca\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105715 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzjfp\" (UniqueName: \"kubernetes.io/projected/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-kube-api-access-tzjfp\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105774 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-datadir\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105798 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-sa-token\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-collector-token\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105849 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-entrypoint\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105897 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-config-openshift-service-cacrt\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.105917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-datadir\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.107309 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-config-openshift-service-cacrt\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.107351 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-config\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.107487 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-trusted-ca\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.107939 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-entrypoint\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.109873 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-tmp\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.110990 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-metrics\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.112020 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-collector-token\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.114586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-collector-syslog-receiver\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.123602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-sa-token\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.124932 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzjfp\" (UniqueName: \"kubernetes.io/projected/1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1-kube-api-access-tzjfp\") pod \"collector-ndnhl\" (UID: \"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1\") " pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.169875 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-ndnhl" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.464629 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db0e77a2-b4bf-46e3-aa53-b5e048f8bb95" path="/var/lib/kubelet/pods/db0e77a2-b4bf-46e3-aa53-b5e048f8bb95/volumes" Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.593449 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-ndnhl"] Nov 27 17:24:19 crc kubenswrapper[4809]: I1127 17:24:19.796592 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-ndnhl" event={"ID":"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1","Type":"ContainerStarted","Data":"9515c733bf11e2cb05f278d5cbec82c539032db5d0ad8c4b9df2cefa4599db10"} Nov 27 17:24:29 crc kubenswrapper[4809]: I1127 17:24:29.855310 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-ndnhl" event={"ID":"1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1","Type":"ContainerStarted","Data":"71943d213ca3a0a1005e87215fd02f3a2efd17876d6281bfd8e089daa614063b"} Nov 27 17:24:29 crc kubenswrapper[4809]: I1127 17:24:29.880142 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-ndnhl" podStartSLOduration=2.280593198 podStartE2EDuration="11.880121576s" podCreationTimestamp="2025-11-27 17:24:18 +0000 UTC" firstStartedPulling="2025-11-27 17:24:19.602637837 +0000 UTC m=+894.875095189" lastFinishedPulling="2025-11-27 17:24:29.202166215 +0000 UTC m=+904.474623567" observedRunningTime="2025-11-27 17:24:29.875406196 +0000 UTC m=+905.147863558" watchObservedRunningTime="2025-11-27 17:24:29.880121576 +0000 UTC m=+905.152578928" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.248306 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p5wd6"] Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.250257 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.259967 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5wd6"] Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.396029 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-catalog-content\") pod \"certified-operators-p5wd6\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.396099 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-utilities\") pod \"certified-operators-p5wd6\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.396200 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkklx\" (UniqueName: \"kubernetes.io/projected/49907e19-c4d8-4c1a-b4aa-5caa94426b85-kube-api-access-vkklx\") pod \"certified-operators-p5wd6\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.497992 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkklx\" (UniqueName: \"kubernetes.io/projected/49907e19-c4d8-4c1a-b4aa-5caa94426b85-kube-api-access-vkklx\") pod \"certified-operators-p5wd6\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.498071 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-catalog-content\") pod \"certified-operators-p5wd6\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.498100 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-utilities\") pod \"certified-operators-p5wd6\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.498615 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-utilities\") pod \"certified-operators-p5wd6\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.498685 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-catalog-content\") pod \"certified-operators-p5wd6\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.525864 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkklx\" (UniqueName: \"kubernetes.io/projected/49907e19-c4d8-4c1a-b4aa-5caa94426b85-kube-api-access-vkklx\") pod \"certified-operators-p5wd6\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.566624 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.827143 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5wd6"] Nov 27 17:24:45 crc kubenswrapper[4809]: I1127 17:24:45.951776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5wd6" event={"ID":"49907e19-c4d8-4c1a-b4aa-5caa94426b85","Type":"ContainerStarted","Data":"f7ce69380cf7899aee2ad854a722e7044ecc7bff9eb71dbc9032167ec4d26820"} Nov 27 17:24:46 crc kubenswrapper[4809]: I1127 17:24:46.958589 4809 generic.go:334] "Generic (PLEG): container finished" podID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerID="4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9" exitCode=0 Nov 27 17:24:46 crc kubenswrapper[4809]: I1127 17:24:46.958636 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5wd6" event={"ID":"49907e19-c4d8-4c1a-b4aa-5caa94426b85","Type":"ContainerDied","Data":"4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9"} Nov 27 17:24:47 crc kubenswrapper[4809]: I1127 17:24:47.968793 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5wd6" event={"ID":"49907e19-c4d8-4c1a-b4aa-5caa94426b85","Type":"ContainerStarted","Data":"6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e"} Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.745442 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.773003 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8zklvf"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.783868 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.792875 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkr5gr"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.798191 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5wd6"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.807777 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pgtgl"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.808068 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pgtgl" podUID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerName="registry-server" containerID="cri-o://05278f23c41f669f29c49c75f01a1253460ed58a948c3c003cfa206717f4f18b" gracePeriod=30 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.818386 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swvpz"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.818700 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-swvpz" podUID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerName="registry-server" containerID="cri-o://e8cfe2308a0d685973241aac606be8f6c64aef23f0bc7403628ed57bd747e79d" gracePeriod=30 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.826858 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-64fv9"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.827019 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" podUID="f2bae280-204f-4f0f-93f5-fd574d9d4b79" containerName="marketplace-operator" containerID="cri-o://f3afb48367da35625b2b19752ebc9d08c76f8857f69cfa18e56a1c71f8ccf2d2" gracePeriod=30 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.830384 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvw4b"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.830567 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fvw4b" podUID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerName="registry-server" containerID="cri-o://c0a5608d1fba6506ac4040e74a1e5d6b93caa2adebd5e150366c2e1153d35527" gracePeriod=30 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.833509 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6pskk"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.833664 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6pskk" podUID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerName="registry-server" containerID="cri-o://818cf7664b2316410f662ce3719970dd8af11acd82bd6b3fa75392625b3b69c4" gracePeriod=30 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.837822 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4x2bl"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.840874 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.846926 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4x2bl"] Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.956774 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmbsp\" (UniqueName: \"kubernetes.io/projected/b3ffdbfe-1e8b-415f-9d97-47a17d84070c-kube-api-access-pmbsp\") pod \"marketplace-operator-79b997595-4x2bl\" (UID: \"b3ffdbfe-1e8b-415f-9d97-47a17d84070c\") " pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.956853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3ffdbfe-1e8b-415f-9d97-47a17d84070c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4x2bl\" (UID: \"b3ffdbfe-1e8b-415f-9d97-47a17d84070c\") " pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.956900 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b3ffdbfe-1e8b-415f-9d97-47a17d84070c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4x2bl\" (UID: \"b3ffdbfe-1e8b-415f-9d97-47a17d84070c\") " pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.982117 4809 generic.go:334] "Generic (PLEG): container finished" podID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerID="6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e" exitCode=0 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.982214 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5wd6" event={"ID":"49907e19-c4d8-4c1a-b4aa-5caa94426b85","Type":"ContainerDied","Data":"6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e"} Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.989317 4809 generic.go:334] "Generic (PLEG): container finished" podID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerID="818cf7664b2316410f662ce3719970dd8af11acd82bd6b3fa75392625b3b69c4" exitCode=0 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.989399 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pskk" event={"ID":"282b8b99-95c9-45b5-9608-f57c52aae9e0","Type":"ContainerDied","Data":"818cf7664b2316410f662ce3719970dd8af11acd82bd6b3fa75392625b3b69c4"} Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.993300 4809 generic.go:334] "Generic (PLEG): container finished" podID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerID="e8cfe2308a0d685973241aac606be8f6c64aef23f0bc7403628ed57bd747e79d" exitCode=0 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.993446 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvpz" event={"ID":"57211ee0-a59b-4c8b-99d0-6854e00b9e53","Type":"ContainerDied","Data":"e8cfe2308a0d685973241aac606be8f6c64aef23f0bc7403628ed57bd747e79d"} Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.997415 4809 generic.go:334] "Generic (PLEG): container finished" podID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerID="c0a5608d1fba6506ac4040e74a1e5d6b93caa2adebd5e150366c2e1153d35527" exitCode=0 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.997479 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvw4b" event={"ID":"f04c3ed8-38c7-445a-b1b7-dfe646d8db71","Type":"ContainerDied","Data":"c0a5608d1fba6506ac4040e74a1e5d6b93caa2adebd5e150366c2e1153d35527"} Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.999009 4809 generic.go:334] "Generic (PLEG): container finished" podID="f2bae280-204f-4f0f-93f5-fd574d9d4b79" containerID="f3afb48367da35625b2b19752ebc9d08c76f8857f69cfa18e56a1c71f8ccf2d2" exitCode=0 Nov 27 17:24:48 crc kubenswrapper[4809]: I1127 17:24:48.999053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" event={"ID":"f2bae280-204f-4f0f-93f5-fd574d9d4b79","Type":"ContainerDied","Data":"f3afb48367da35625b2b19752ebc9d08c76f8857f69cfa18e56a1c71f8ccf2d2"} Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.002439 4809 generic.go:334] "Generic (PLEG): container finished" podID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerID="05278f23c41f669f29c49c75f01a1253460ed58a948c3c003cfa206717f4f18b" exitCode=0 Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.002466 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgtgl" event={"ID":"ceaa3143-e2dc-40fd-8100-253811eeb357","Type":"ContainerDied","Data":"05278f23c41f669f29c49c75f01a1253460ed58a948c3c003cfa206717f4f18b"} Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.059532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmbsp\" (UniqueName: \"kubernetes.io/projected/b3ffdbfe-1e8b-415f-9d97-47a17d84070c-kube-api-access-pmbsp\") pod \"marketplace-operator-79b997595-4x2bl\" (UID: \"b3ffdbfe-1e8b-415f-9d97-47a17d84070c\") " pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.059602 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3ffdbfe-1e8b-415f-9d97-47a17d84070c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4x2bl\" (UID: \"b3ffdbfe-1e8b-415f-9d97-47a17d84070c\") " pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.060588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b3ffdbfe-1e8b-415f-9d97-47a17d84070c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4x2bl\" (UID: \"b3ffdbfe-1e8b-415f-9d97-47a17d84070c\") " pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.061364 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3ffdbfe-1e8b-415f-9d97-47a17d84070c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4x2bl\" (UID: \"b3ffdbfe-1e8b-415f-9d97-47a17d84070c\") " pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.068089 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b3ffdbfe-1e8b-415f-9d97-47a17d84070c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4x2bl\" (UID: \"b3ffdbfe-1e8b-415f-9d97-47a17d84070c\") " pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.082535 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmbsp\" (UniqueName: \"kubernetes.io/projected/b3ffdbfe-1e8b-415f-9d97-47a17d84070c-kube-api-access-pmbsp\") pod \"marketplace-operator-79b997595-4x2bl\" (UID: \"b3ffdbfe-1e8b-415f-9d97-47a17d84070c\") " pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.165788 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.353336 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.361729 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.364184 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.369910 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.387649 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.466199 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-catalog-content\") pod \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.466269 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmnwt\" (UniqueName: \"kubernetes.io/projected/ceaa3143-e2dc-40fd-8100-253811eeb357-kube-api-access-hmnwt\") pod \"ceaa3143-e2dc-40fd-8100-253811eeb357\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.466307 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-utilities\") pod \"282b8b99-95c9-45b5-9608-f57c52aae9e0\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.466346 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-utilities\") pod \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.466394 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8szx\" (UniqueName: \"kubernetes.io/projected/f2bae280-204f-4f0f-93f5-fd574d9d4b79-kube-api-access-z8szx\") pod \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.466465 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-utilities\") pod \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.466558 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-catalog-content\") pod \"282b8b99-95c9-45b5-9608-f57c52aae9e0\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.466596 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qsxn\" (UniqueName: \"kubernetes.io/projected/57211ee0-a59b-4c8b-99d0-6854e00b9e53-kube-api-access-7qsxn\") pod \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\" (UID: \"57211ee0-a59b-4c8b-99d0-6854e00b9e53\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.466646 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtq8v\" (UniqueName: \"kubernetes.io/projected/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-kube-api-access-rtq8v\") pod \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.468286 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-catalog-content\") pod \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\" (UID: \"f04c3ed8-38c7-445a-b1b7-dfe646d8db71\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.468333 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-operator-metrics\") pod \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.468401 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-catalog-content\") pod \"ceaa3143-e2dc-40fd-8100-253811eeb357\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.469503 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-utilities" (OuterVolumeSpecName: "utilities") pod "f04c3ed8-38c7-445a-b1b7-dfe646d8db71" (UID: "f04c3ed8-38c7-445a-b1b7-dfe646d8db71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.469564 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-utilities" (OuterVolumeSpecName: "utilities") pod "282b8b99-95c9-45b5-9608-f57c52aae9e0" (UID: "282b8b99-95c9-45b5-9608-f57c52aae9e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.469599 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-utilities" (OuterVolumeSpecName: "utilities") pod "57211ee0-a59b-4c8b-99d0-6854e00b9e53" (UID: "57211ee0-a59b-4c8b-99d0-6854e00b9e53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.469938 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/282b8b99-95c9-45b5-9608-f57c52aae9e0-kube-api-access-qpmw2\") pod \"282b8b99-95c9-45b5-9608-f57c52aae9e0\" (UID: \"282b8b99-95c9-45b5-9608-f57c52aae9e0\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.470111 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-trusted-ca\") pod \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\" (UID: \"f2bae280-204f-4f0f-93f5-fd574d9d4b79\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.470160 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-utilities\") pod \"ceaa3143-e2dc-40fd-8100-253811eeb357\" (UID: \"ceaa3143-e2dc-40fd-8100-253811eeb357\") " Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.471246 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.471271 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.471286 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.477135 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57adfa91-4c88-4dc8-a92c-110ef12b330c" path="/var/lib/kubelet/pods/57adfa91-4c88-4dc8-a92c-110ef12b330c/volumes" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.477185 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57211ee0-a59b-4c8b-99d0-6854e00b9e53-kube-api-access-7qsxn" (OuterVolumeSpecName: "kube-api-access-7qsxn") pod "57211ee0-a59b-4c8b-99d0-6854e00b9e53" (UID: "57211ee0-a59b-4c8b-99d0-6854e00b9e53"). InnerVolumeSpecName "kube-api-access-7qsxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.478120 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f2bae280-204f-4f0f-93f5-fd574d9d4b79" (UID: "f2bae280-204f-4f0f-93f5-fd574d9d4b79"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.478341 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2bae280-204f-4f0f-93f5-fd574d9d4b79-kube-api-access-z8szx" (OuterVolumeSpecName: "kube-api-access-z8szx") pod "f2bae280-204f-4f0f-93f5-fd574d9d4b79" (UID: "f2bae280-204f-4f0f-93f5-fd574d9d4b79"). InnerVolumeSpecName "kube-api-access-z8szx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.478486 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88f51563-f860-48b1-b0e5-f5c11bfc78e5" path="/var/lib/kubelet/pods/88f51563-f860-48b1-b0e5-f5c11bfc78e5/volumes" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.479245 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-kube-api-access-rtq8v" (OuterVolumeSpecName: "kube-api-access-rtq8v") pod "f04c3ed8-38c7-445a-b1b7-dfe646d8db71" (UID: "f04c3ed8-38c7-445a-b1b7-dfe646d8db71"). InnerVolumeSpecName "kube-api-access-rtq8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.479304 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceaa3143-e2dc-40fd-8100-253811eeb357-kube-api-access-hmnwt" (OuterVolumeSpecName: "kube-api-access-hmnwt") pod "ceaa3143-e2dc-40fd-8100-253811eeb357" (UID: "ceaa3143-e2dc-40fd-8100-253811eeb357"). InnerVolumeSpecName "kube-api-access-hmnwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.486382 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/282b8b99-95c9-45b5-9608-f57c52aae9e0-kube-api-access-qpmw2" (OuterVolumeSpecName: "kube-api-access-qpmw2") pod "282b8b99-95c9-45b5-9608-f57c52aae9e0" (UID: "282b8b99-95c9-45b5-9608-f57c52aae9e0"). InnerVolumeSpecName "kube-api-access-qpmw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.486405 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-utilities" (OuterVolumeSpecName: "utilities") pod "ceaa3143-e2dc-40fd-8100-253811eeb357" (UID: "ceaa3143-e2dc-40fd-8100-253811eeb357"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.486728 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f2bae280-204f-4f0f-93f5-fd574d9d4b79" (UID: "f2bae280-204f-4f0f-93f5-fd574d9d4b79"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.504295 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4x2bl"] Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.510160 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f04c3ed8-38c7-445a-b1b7-dfe646d8db71" (UID: "f04c3ed8-38c7-445a-b1b7-dfe646d8db71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: W1127 17:24:49.510688 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3ffdbfe_1e8b_415f_9d97_47a17d84070c.slice/crio-8881bd9d221cd019edf997fd80d4afe4806c1066f010659e9a98c74d14e15454 WatchSource:0}: Error finding container 8881bd9d221cd019edf997fd80d4afe4806c1066f010659e9a98c74d14e15454: Status 404 returned error can't find the container with id 8881bd9d221cd019edf997fd80d4afe4806c1066f010659e9a98c74d14e15454 Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.543658 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57211ee0-a59b-4c8b-99d0-6854e00b9e53" (UID: "57211ee0-a59b-4c8b-99d0-6854e00b9e53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.558922 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ceaa3143-e2dc-40fd-8100-253811eeb357" (UID: "ceaa3143-e2dc-40fd-8100-253811eeb357"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572430 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57211ee0-a59b-4c8b-99d0-6854e00b9e53-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572465 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmnwt\" (UniqueName: \"kubernetes.io/projected/ceaa3143-e2dc-40fd-8100-253811eeb357-kube-api-access-hmnwt\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572479 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8szx\" (UniqueName: \"kubernetes.io/projected/f2bae280-204f-4f0f-93f5-fd574d9d4b79-kube-api-access-z8szx\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572493 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qsxn\" (UniqueName: \"kubernetes.io/projected/57211ee0-a59b-4c8b-99d0-6854e00b9e53-kube-api-access-7qsxn\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572505 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtq8v\" (UniqueName: \"kubernetes.io/projected/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-kube-api-access-rtq8v\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572517 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c3ed8-38c7-445a-b1b7-dfe646d8db71-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572527 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572537 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572547 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/282b8b99-95c9-45b5-9608-f57c52aae9e0-kube-api-access-qpmw2\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572558 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2bae280-204f-4f0f-93f5-fd574d9d4b79-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.572568 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceaa3143-e2dc-40fd-8100-253811eeb357-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.626307 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "282b8b99-95c9-45b5-9608-f57c52aae9e0" (UID: "282b8b99-95c9-45b5-9608-f57c52aae9e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:49 crc kubenswrapper[4809]: I1127 17:24:49.673872 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/282b8b99-95c9-45b5-9608-f57c52aae9e0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.011037 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5wd6" event={"ID":"49907e19-c4d8-4c1a-b4aa-5caa94426b85","Type":"ContainerStarted","Data":"b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501"} Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.011259 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p5wd6" podUID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerName="registry-server" containerID="cri-o://b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501" gracePeriod=30 Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.014420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6pskk" event={"ID":"282b8b99-95c9-45b5-9608-f57c52aae9e0","Type":"ContainerDied","Data":"5eacc89d2dddc19cfa197f02ab79bf76c4b15e27e3e03c5cacee332cd36fe463"} Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.014468 4809 scope.go:117] "RemoveContainer" containerID="818cf7664b2316410f662ce3719970dd8af11acd82bd6b3fa75392625b3b69c4" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.014621 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6pskk" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.025450 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swvpz" event={"ID":"57211ee0-a59b-4c8b-99d0-6854e00b9e53","Type":"ContainerDied","Data":"73e2ca9389ce5c5c71530acb183a13806133870e028765585d894ac036217af3"} Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.025612 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swvpz" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.036367 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" event={"ID":"b3ffdbfe-1e8b-415f-9d97-47a17d84070c","Type":"ContainerStarted","Data":"5cffb0b1248f024e03247dc8d7620a6b48e877539e59eb9495f0b811b40736f4"} Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.036433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" event={"ID":"b3ffdbfe-1e8b-415f-9d97-47a17d84070c","Type":"ContainerStarted","Data":"8881bd9d221cd019edf997fd80d4afe4806c1066f010659e9a98c74d14e15454"} Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.036929 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.043690 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.044496 4809 scope.go:117] "RemoveContainer" containerID="70cee96759554b5537cf9fdf0e16948f95bac4219620978c5c1fc4c2f289e86a" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.044579 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvw4b" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.044599 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvw4b" event={"ID":"f04c3ed8-38c7-445a-b1b7-dfe646d8db71","Type":"ContainerDied","Data":"b2ca0358e9473632226acfc01005a08dad1d4e0832a867eeff2ab5d61a9a1107"} Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.046424 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p5wd6" podStartSLOduration=2.4389356380000002 podStartE2EDuration="5.046412377s" podCreationTimestamp="2025-11-27 17:24:45 +0000 UTC" firstStartedPulling="2025-11-27 17:24:46.961647724 +0000 UTC m=+922.234105116" lastFinishedPulling="2025-11-27 17:24:49.569124503 +0000 UTC m=+924.841581855" observedRunningTime="2025-11-27 17:24:50.040102775 +0000 UTC m=+925.312560127" watchObservedRunningTime="2025-11-27 17:24:50.046412377 +0000 UTC m=+925.318869729" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.047054 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.047874 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-64fv9" event={"ID":"f2bae280-204f-4f0f-93f5-fd574d9d4b79","Type":"ContainerDied","Data":"33c084717ecee0daeb906fd16edeac01ba603b9a8692bd529b2e29abc8506285"} Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.051475 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgtgl" event={"ID":"ceaa3143-e2dc-40fd-8100-253811eeb357","Type":"ContainerDied","Data":"b757815544d05d9b7f2872f8a330c1f600016a3da560fe9325558f8489584f72"} Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.051552 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgtgl" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.068365 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4x2bl" podStartSLOduration=2.068339277 podStartE2EDuration="2.068339277s" podCreationTimestamp="2025-11-27 17:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:24:50.06624299 +0000 UTC m=+925.338700342" watchObservedRunningTime="2025-11-27 17:24:50.068339277 +0000 UTC m=+925.340796629" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.089786 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6pskk"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.091995 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6pskk"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.113562 4809 scope.go:117] "RemoveContainer" containerID="100db1106b92fef5e567bd4b16ee21181be84b474fd88a2636a2baf6d1c6222e" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.129508 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-64fv9"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.154344 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-64fv9"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.161111 4809 scope.go:117] "RemoveContainer" containerID="e8cfe2308a0d685973241aac606be8f6c64aef23f0bc7403628ed57bd747e79d" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.198916 4809 scope.go:117] "RemoveContainer" containerID="5c459bb53c76e8f422b63e4b290c124a84aa77ee269dca530d24c6f01585f3e2" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.203970 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvw4b"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.216177 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvw4b"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.227917 4809 scope.go:117] "RemoveContainer" containerID="42e863447013efda8e741ebac2812c74c25b3ba5dd50d509aad99f6d4443019e" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.228988 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swvpz"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.236320 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-swvpz"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.240111 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pgtgl"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.243480 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pgtgl"] Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.251366 4809 scope.go:117] "RemoveContainer" containerID="c0a5608d1fba6506ac4040e74a1e5d6b93caa2adebd5e150366c2e1153d35527" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.270131 4809 scope.go:117] "RemoveContainer" containerID="2ad80a4b4ea958ee1d3328c427830df599c322f09caa03c95cc74b8edfc4f62d" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.289880 4809 scope.go:117] "RemoveContainer" containerID="116203be209a4a802ed3f364fe944c6f2b96f645dbdf347e9a6018351dc4c4e6" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.305348 4809 scope.go:117] "RemoveContainer" containerID="f3afb48367da35625b2b19752ebc9d08c76f8857f69cfa18e56a1c71f8ccf2d2" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.321532 4809 scope.go:117] "RemoveContainer" containerID="05278f23c41f669f29c49c75f01a1253460ed58a948c3c003cfa206717f4f18b" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.347236 4809 scope.go:117] "RemoveContainer" containerID="fb3871a99fc82b18aca4345cb5e6d6a47649302dd57bf9409aa9dea99a8ee064" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.367919 4809 scope.go:117] "RemoveContainer" containerID="27823b4e69930199d9cb28f54256cd3eef8f1f7ea84e8613fc1db3178a339ec1" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.405716 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p5wd6_49907e19-c4d8-4c1a-b4aa-5caa94426b85/registry-server/0.log" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.406445 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.487198 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-utilities\") pod \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.487508 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-catalog-content\") pod \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.487562 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkklx\" (UniqueName: \"kubernetes.io/projected/49907e19-c4d8-4c1a-b4aa-5caa94426b85-kube-api-access-vkklx\") pod \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\" (UID: \"49907e19-c4d8-4c1a-b4aa-5caa94426b85\") " Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.488329 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-utilities" (OuterVolumeSpecName: "utilities") pod "49907e19-c4d8-4c1a-b4aa-5caa94426b85" (UID: "49907e19-c4d8-4c1a-b4aa-5caa94426b85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.489187 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.495774 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49907e19-c4d8-4c1a-b4aa-5caa94426b85-kube-api-access-vkklx" (OuterVolumeSpecName: "kube-api-access-vkklx") pod "49907e19-c4d8-4c1a-b4aa-5caa94426b85" (UID: "49907e19-c4d8-4c1a-b4aa-5caa94426b85"). InnerVolumeSpecName "kube-api-access-vkklx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.534979 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49907e19-c4d8-4c1a-b4aa-5caa94426b85" (UID: "49907e19-c4d8-4c1a-b4aa-5caa94426b85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.590193 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49907e19-c4d8-4c1a-b4aa-5caa94426b85-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:50 crc kubenswrapper[4809]: I1127 17:24:50.591094 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkklx\" (UniqueName: \"kubernetes.io/projected/49907e19-c4d8-4c1a-b4aa-5caa94426b85-kube-api-access-vkklx\") on node \"crc\" DevicePath \"\"" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.026985 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-28jvg"] Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027285 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027303 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027320 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027328 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027342 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027351 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027360 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027369 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027375 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027382 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027395 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027404 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027412 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027419 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027430 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027438 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027447 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2bae280-204f-4f0f-93f5-fd574d9d4b79" containerName="marketplace-operator" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027454 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2bae280-204f-4f0f-93f5-fd574d9d4b79" containerName="marketplace-operator" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027464 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027472 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027484 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027491 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027500 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027507 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerName="extract-utilities" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027515 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027526 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027538 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027545 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerName="extract-content" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027555 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027562 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.027574 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027581 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027699 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2bae280-204f-4f0f-93f5-fd574d9d4b79" containerName="marketplace-operator" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027715 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027726 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027757 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceaa3143-e2dc-40fd-8100-253811eeb357" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027771 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.027784 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="282b8b99-95c9-45b5-9608-f57c52aae9e0" containerName="registry-server" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.028785 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.032505 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.037262 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-28jvg"] Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.069789 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p5wd6_49907e19-c4d8-4c1a-b4aa-5caa94426b85/registry-server/0.log" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.070403 4809 generic.go:334] "Generic (PLEG): container finished" podID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" containerID="b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501" exitCode=1 Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.070459 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5wd6" event={"ID":"49907e19-c4d8-4c1a-b4aa-5caa94426b85","Type":"ContainerDied","Data":"b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501"} Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.070493 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5wd6" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.070510 4809 scope.go:117] "RemoveContainer" containerID="b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.070493 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5wd6" event={"ID":"49907e19-c4d8-4c1a-b4aa-5caa94426b85","Type":"ContainerDied","Data":"f7ce69380cf7899aee2ad854a722e7044ecc7bff9eb71dbc9032167ec4d26820"} Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.087931 4809 scope.go:117] "RemoveContainer" containerID="6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.097036 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-utilities\") pod \"community-operators-28jvg\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.097100 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-catalog-content\") pod \"community-operators-28jvg\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.097145 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpp5s\" (UniqueName: \"kubernetes.io/projected/492f5beb-f94c-4510-a07b-306ef7d369b8-kube-api-access-zpp5s\") pod \"community-operators-28jvg\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.097210 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5wd6"] Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.100626 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p5wd6"] Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.123418 4809 scope.go:117] "RemoveContainer" containerID="4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.136383 4809 scope.go:117] "RemoveContainer" containerID="b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.137115 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501\": container with ID starting with b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501 not found: ID does not exist" containerID="b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.137184 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501"} err="failed to get container status \"b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501\": rpc error: code = NotFound desc = could not find container \"b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501\": container with ID starting with b666153eaab6f0c4a19b063166f9f78cc7556d1744771e85e946a58aa933b501 not found: ID does not exist" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.137215 4809 scope.go:117] "RemoveContainer" containerID="6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.137661 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e\": container with ID starting with 6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e not found: ID does not exist" containerID="6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.137708 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e"} err="failed to get container status \"6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e\": rpc error: code = NotFound desc = could not find container \"6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e\": container with ID starting with 6d93fde00d07b9b02ee89c835fc43db2d72ca946b2864bb78f2a2e0c0d288e2e not found: ID does not exist" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.137730 4809 scope.go:117] "RemoveContainer" containerID="4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9" Nov 27 17:24:51 crc kubenswrapper[4809]: E1127 17:24:51.138093 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9\": container with ID starting with 4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9 not found: ID does not exist" containerID="4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.138143 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9"} err="failed to get container status \"4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9\": rpc error: code = NotFound desc = could not find container \"4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9\": container with ID starting with 4dd183c5d3297b9d70f1c4b7b58f76cc23c77a4ba173ce07f7ff8a551e4cbda9 not found: ID does not exist" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.197997 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-catalog-content\") pod \"community-operators-28jvg\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.198109 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpp5s\" (UniqueName: \"kubernetes.io/projected/492f5beb-f94c-4510-a07b-306ef7d369b8-kube-api-access-zpp5s\") pod \"community-operators-28jvg\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.198182 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-utilities\") pod \"community-operators-28jvg\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.198568 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-catalog-content\") pod \"community-operators-28jvg\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.198591 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-utilities\") pod \"community-operators-28jvg\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.216520 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpp5s\" (UniqueName: \"kubernetes.io/projected/492f5beb-f94c-4510-a07b-306ef7d369b8-kube-api-access-zpp5s\") pod \"community-operators-28jvg\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.344888 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.429144 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8r4bc"] Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.432939 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.450388 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8r4bc"] Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.470285 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="282b8b99-95c9-45b5-9608-f57c52aae9e0" path="/var/lib/kubelet/pods/282b8b99-95c9-45b5-9608-f57c52aae9e0/volumes" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.471539 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49907e19-c4d8-4c1a-b4aa-5caa94426b85" path="/var/lib/kubelet/pods/49907e19-c4d8-4c1a-b4aa-5caa94426b85/volumes" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.472699 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57211ee0-a59b-4c8b-99d0-6854e00b9e53" path="/var/lib/kubelet/pods/57211ee0-a59b-4c8b-99d0-6854e00b9e53/volumes" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.475128 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceaa3143-e2dc-40fd-8100-253811eeb357" path="/var/lib/kubelet/pods/ceaa3143-e2dc-40fd-8100-253811eeb357/volumes" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.476537 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f04c3ed8-38c7-445a-b1b7-dfe646d8db71" path="/var/lib/kubelet/pods/f04c3ed8-38c7-445a-b1b7-dfe646d8db71/volumes" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.478224 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2bae280-204f-4f0f-93f5-fd574d9d4b79" path="/var/lib/kubelet/pods/f2bae280-204f-4f0f-93f5-fd574d9d4b79/volumes" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.502877 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-catalog-content\") pod \"community-operators-8r4bc\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.503558 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4vgd\" (UniqueName: \"kubernetes.io/projected/1b5febb2-32de-4c4a-a8fb-1168893feff9-kube-api-access-k4vgd\") pod \"community-operators-8r4bc\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.503812 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-utilities\") pod \"community-operators-8r4bc\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.605450 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-utilities\") pod \"community-operators-8r4bc\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.605595 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-catalog-content\") pod \"community-operators-8r4bc\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.605712 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4vgd\" (UniqueName: \"kubernetes.io/projected/1b5febb2-32de-4c4a-a8fb-1168893feff9-kube-api-access-k4vgd\") pod \"community-operators-8r4bc\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.606059 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-utilities\") pod \"community-operators-8r4bc\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.606066 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-catalog-content\") pod \"community-operators-8r4bc\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.628658 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4vgd\" (UniqueName: \"kubernetes.io/projected/1b5febb2-32de-4c4a-a8fb-1168893feff9-kube-api-access-k4vgd\") pod \"community-operators-8r4bc\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.754243 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:24:51 crc kubenswrapper[4809]: I1127 17:24:51.777974 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-28jvg"] Nov 27 17:24:51 crc kubenswrapper[4809]: W1127 17:24:51.778037 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod492f5beb_f94c_4510_a07b_306ef7d369b8.slice/crio-ea2084e6d4c5e4c504dd08683eee6ee3b3c2686b792df2955a8480e2442e3670 WatchSource:0}: Error finding container ea2084e6d4c5e4c504dd08683eee6ee3b3c2686b792df2955a8480e2442e3670: Status 404 returned error can't find the container with id ea2084e6d4c5e4c504dd08683eee6ee3b3c2686b792df2955a8480e2442e3670 Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.081967 4809 generic.go:334] "Generic (PLEG): container finished" podID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerID="62db8d1d2f0dea49b01f5d278ddbe593a19f389b050682dd0aea3f4ef9ffeb39" exitCode=0 Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.082048 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28jvg" event={"ID":"492f5beb-f94c-4510-a07b-306ef7d369b8","Type":"ContainerDied","Data":"62db8d1d2f0dea49b01f5d278ddbe593a19f389b050682dd0aea3f4ef9ffeb39"} Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.082111 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28jvg" event={"ID":"492f5beb-f94c-4510-a07b-306ef7d369b8","Type":"ContainerStarted","Data":"ea2084e6d4c5e4c504dd08683eee6ee3b3c2686b792df2955a8480e2442e3670"} Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.158358 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8r4bc"] Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.425594 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hk5rd"] Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.427279 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.429425 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.433423 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hk5rd"] Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.518507 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh44b\" (UniqueName: \"kubernetes.io/projected/24a4f327-f4ee-45dd-938e-4de8c52b2e76-kube-api-access-mh44b\") pod \"redhat-marketplace-hk5rd\" (UID: \"24a4f327-f4ee-45dd-938e-4de8c52b2e76\") " pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.518567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a4f327-f4ee-45dd-938e-4de8c52b2e76-utilities\") pod \"redhat-marketplace-hk5rd\" (UID: \"24a4f327-f4ee-45dd-938e-4de8c52b2e76\") " pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.518594 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a4f327-f4ee-45dd-938e-4de8c52b2e76-catalog-content\") pod \"redhat-marketplace-hk5rd\" (UID: \"24a4f327-f4ee-45dd-938e-4de8c52b2e76\") " pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.620475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh44b\" (UniqueName: \"kubernetes.io/projected/24a4f327-f4ee-45dd-938e-4de8c52b2e76-kube-api-access-mh44b\") pod \"redhat-marketplace-hk5rd\" (UID: \"24a4f327-f4ee-45dd-938e-4de8c52b2e76\") " pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.621036 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a4f327-f4ee-45dd-938e-4de8c52b2e76-utilities\") pod \"redhat-marketplace-hk5rd\" (UID: \"24a4f327-f4ee-45dd-938e-4de8c52b2e76\") " pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.621092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a4f327-f4ee-45dd-938e-4de8c52b2e76-catalog-content\") pod \"redhat-marketplace-hk5rd\" (UID: \"24a4f327-f4ee-45dd-938e-4de8c52b2e76\") " pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.621539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a4f327-f4ee-45dd-938e-4de8c52b2e76-utilities\") pod \"redhat-marketplace-hk5rd\" (UID: \"24a4f327-f4ee-45dd-938e-4de8c52b2e76\") " pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.621863 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a4f327-f4ee-45dd-938e-4de8c52b2e76-catalog-content\") pod \"redhat-marketplace-hk5rd\" (UID: \"24a4f327-f4ee-45dd-938e-4de8c52b2e76\") " pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.642933 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh44b\" (UniqueName: \"kubernetes.io/projected/24a4f327-f4ee-45dd-938e-4de8c52b2e76-kube-api-access-mh44b\") pod \"redhat-marketplace-hk5rd\" (UID: \"24a4f327-f4ee-45dd-938e-4de8c52b2e76\") " pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.743525 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.829864 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hd5db"] Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.831167 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.842428 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd5db"] Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.927033 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5jv2\" (UniqueName: \"kubernetes.io/projected/7713e7ad-760d-4807-b2cf-17df91b637c0-kube-api-access-t5jv2\") pod \"redhat-marketplace-hd5db\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.927117 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-utilities\") pod \"redhat-marketplace-hd5db\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:52 crc kubenswrapper[4809]: I1127 17:24:52.927233 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-catalog-content\") pod \"redhat-marketplace-hd5db\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.028968 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-catalog-content\") pod \"redhat-marketplace-hd5db\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.029122 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5jv2\" (UniqueName: \"kubernetes.io/projected/7713e7ad-760d-4807-b2cf-17df91b637c0-kube-api-access-t5jv2\") pod \"redhat-marketplace-hd5db\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.029145 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-utilities\") pod \"redhat-marketplace-hd5db\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.029514 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-catalog-content\") pod \"redhat-marketplace-hd5db\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.029770 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-utilities\") pod \"redhat-marketplace-hd5db\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.046754 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5jv2\" (UniqueName: \"kubernetes.io/projected/7713e7ad-760d-4807-b2cf-17df91b637c0-kube-api-access-t5jv2\") pod \"redhat-marketplace-hd5db\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.091183 4809 generic.go:334] "Generic (PLEG): container finished" podID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerID="aa22dcf997faf7d8f05c1279b6e9c13a17615b401f95b58679d418ad719cccb5" exitCode=0 Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.091275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8r4bc" event={"ID":"1b5febb2-32de-4c4a-a8fb-1168893feff9","Type":"ContainerDied","Data":"aa22dcf997faf7d8f05c1279b6e9c13a17615b401f95b58679d418ad719cccb5"} Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.091311 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8r4bc" event={"ID":"1b5febb2-32de-4c4a-a8fb-1168893feff9","Type":"ContainerStarted","Data":"c7722dba898a668fe4c982c933c9a44faab122f3501dcace610119ce4e39797d"} Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.154886 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hk5rd"] Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.155295 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.563495 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd5db"] Nov 27 17:24:53 crc kubenswrapper[4809]: W1127 17:24:53.567764 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7713e7ad_760d_4807_b2cf_17df91b637c0.slice/crio-52d10367defdf2ca54d855baddb743a6d7031b8079bc3020f346ec2b5a3d6b7d WatchSource:0}: Error finding container 52d10367defdf2ca54d855baddb743a6d7031b8079bc3020f346ec2b5a3d6b7d: Status 404 returned error can't find the container with id 52d10367defdf2ca54d855baddb743a6d7031b8079bc3020f346ec2b5a3d6b7d Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.823262 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8zfdv"] Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.824405 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.828831 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.833270 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zfdv"] Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.940395 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p44f5\" (UniqueName: \"kubernetes.io/projected/ac0a6dfd-901d-47b5-aa9e-44e040d13e75-kube-api-access-p44f5\") pod \"redhat-operators-8zfdv\" (UID: \"ac0a6dfd-901d-47b5-aa9e-44e040d13e75\") " pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.940450 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac0a6dfd-901d-47b5-aa9e-44e040d13e75-catalog-content\") pod \"redhat-operators-8zfdv\" (UID: \"ac0a6dfd-901d-47b5-aa9e-44e040d13e75\") " pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:53 crc kubenswrapper[4809]: I1127 17:24:53.940486 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac0a6dfd-901d-47b5-aa9e-44e040d13e75-utilities\") pod \"redhat-operators-8zfdv\" (UID: \"ac0a6dfd-901d-47b5-aa9e-44e040d13e75\") " pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.042142 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p44f5\" (UniqueName: \"kubernetes.io/projected/ac0a6dfd-901d-47b5-aa9e-44e040d13e75-kube-api-access-p44f5\") pod \"redhat-operators-8zfdv\" (UID: \"ac0a6dfd-901d-47b5-aa9e-44e040d13e75\") " pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.042212 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac0a6dfd-901d-47b5-aa9e-44e040d13e75-catalog-content\") pod \"redhat-operators-8zfdv\" (UID: \"ac0a6dfd-901d-47b5-aa9e-44e040d13e75\") " pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.042247 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac0a6dfd-901d-47b5-aa9e-44e040d13e75-utilities\") pod \"redhat-operators-8zfdv\" (UID: \"ac0a6dfd-901d-47b5-aa9e-44e040d13e75\") " pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.042905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac0a6dfd-901d-47b5-aa9e-44e040d13e75-utilities\") pod \"redhat-operators-8zfdv\" (UID: \"ac0a6dfd-901d-47b5-aa9e-44e040d13e75\") " pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.042959 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac0a6dfd-901d-47b5-aa9e-44e040d13e75-catalog-content\") pod \"redhat-operators-8zfdv\" (UID: \"ac0a6dfd-901d-47b5-aa9e-44e040d13e75\") " pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.062579 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p44f5\" (UniqueName: \"kubernetes.io/projected/ac0a6dfd-901d-47b5-aa9e-44e040d13e75-kube-api-access-p44f5\") pod \"redhat-operators-8zfdv\" (UID: \"ac0a6dfd-901d-47b5-aa9e-44e040d13e75\") " pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.099398 4809 generic.go:334] "Generic (PLEG): container finished" podID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerID="c0193f4086a576d33312924f176de8c41d1706b5d0e03cfa7bd9efa79794c915" exitCode=0 Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.099488 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd5db" event={"ID":"7713e7ad-760d-4807-b2cf-17df91b637c0","Type":"ContainerDied","Data":"c0193f4086a576d33312924f176de8c41d1706b5d0e03cfa7bd9efa79794c915"} Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.099810 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd5db" event={"ID":"7713e7ad-760d-4807-b2cf-17df91b637c0","Type":"ContainerStarted","Data":"52d10367defdf2ca54d855baddb743a6d7031b8079bc3020f346ec2b5a3d6b7d"} Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.106468 4809 generic.go:334] "Generic (PLEG): container finished" podID="24a4f327-f4ee-45dd-938e-4de8c52b2e76" containerID="48d48ee68b8a00e62ee4afb30fe8cf3c2a7eb97facb6c249f484b235c120f90c" exitCode=0 Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.106557 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk5rd" event={"ID":"24a4f327-f4ee-45dd-938e-4de8c52b2e76","Type":"ContainerDied","Data":"48d48ee68b8a00e62ee4afb30fe8cf3c2a7eb97facb6c249f484b235c120f90c"} Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.106586 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk5rd" event={"ID":"24a4f327-f4ee-45dd-938e-4de8c52b2e76","Type":"ContainerStarted","Data":"a75ca1dec0aceb87bb973b22a9d53ce4bb6f1bc3d12fec9d69237a087756602b"} Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.111791 4809 generic.go:334] "Generic (PLEG): container finished" podID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerID="16b24f37185280255bcacedf08aaec7ce277eccc2b2b0c3e54a89599bc98fbe4" exitCode=0 Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.111831 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28jvg" event={"ID":"492f5beb-f94c-4510-a07b-306ef7d369b8","Type":"ContainerDied","Data":"16b24f37185280255bcacedf08aaec7ce277eccc2b2b0c3e54a89599bc98fbe4"} Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.139007 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:24:54 crc kubenswrapper[4809]: I1127 17:24:54.533168 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zfdv"] Nov 27 17:24:54 crc kubenswrapper[4809]: W1127 17:24:54.543010 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac0a6dfd_901d_47b5_aa9e_44e040d13e75.slice/crio-63ccccd6840e92caef921dfa99170b04e5b523c91932c99ff6efeac7ae9915e8 WatchSource:0}: Error finding container 63ccccd6840e92caef921dfa99170b04e5b523c91932c99ff6efeac7ae9915e8: Status 404 returned error can't find the container with id 63ccccd6840e92caef921dfa99170b04e5b523c91932c99ff6efeac7ae9915e8 Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.120425 4809 generic.go:334] "Generic (PLEG): container finished" podID="ac0a6dfd-901d-47b5-aa9e-44e040d13e75" containerID="2c1a45bdf3d90fd87f32d755a22e9a1213dcf494295c3c3016035d5da2c9cea4" exitCode=0 Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.120478 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zfdv" event={"ID":"ac0a6dfd-901d-47b5-aa9e-44e040d13e75","Type":"ContainerDied","Data":"2c1a45bdf3d90fd87f32d755a22e9a1213dcf494295c3c3016035d5da2c9cea4"} Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.120535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zfdv" event={"ID":"ac0a6dfd-901d-47b5-aa9e-44e040d13e75","Type":"ContainerStarted","Data":"63ccccd6840e92caef921dfa99170b04e5b523c91932c99ff6efeac7ae9915e8"} Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.123375 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28jvg" event={"ID":"492f5beb-f94c-4510-a07b-306ef7d369b8","Type":"ContainerStarted","Data":"4a4313aa659ffaca6983a372b89dc53a5631a96fe32af46ce252e3d35fc6392c"} Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.125525 4809 generic.go:334] "Generic (PLEG): container finished" podID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerID="47686c2ce8e17cd660aee93f35a87b1aba8fcddbbe86ce0cfd6c4f1758cbb684" exitCode=0 Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.125568 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8r4bc" event={"ID":"1b5febb2-32de-4c4a-a8fb-1168893feff9","Type":"ContainerDied","Data":"47686c2ce8e17cd660aee93f35a87b1aba8fcddbbe86ce0cfd6c4f1758cbb684"} Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.229725 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-28jvg" podStartSLOduration=1.6075219 podStartE2EDuration="4.229699611s" podCreationTimestamp="2025-11-27 17:24:51 +0000 UTC" firstStartedPulling="2025-11-27 17:24:52.084024412 +0000 UTC m=+927.356481764" lastFinishedPulling="2025-11-27 17:24:54.706202123 +0000 UTC m=+929.978659475" observedRunningTime="2025-11-27 17:24:55.194002015 +0000 UTC m=+930.466459367" watchObservedRunningTime="2025-11-27 17:24:55.229699611 +0000 UTC m=+930.502156953" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.230626 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ggq4r"] Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.231884 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.233921 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.247355 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ggq4r"] Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.363096 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0-utilities\") pod \"certified-operators-ggq4r\" (UID: \"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0\") " pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.363202 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0-catalog-content\") pod \"certified-operators-ggq4r\" (UID: \"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0\") " pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.363294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxz9d\" (UniqueName: \"kubernetes.io/projected/aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0-kube-api-access-wxz9d\") pod \"certified-operators-ggq4r\" (UID: \"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0\") " pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.464197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0-catalog-content\") pod \"certified-operators-ggq4r\" (UID: \"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0\") " pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.464261 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxz9d\" (UniqueName: \"kubernetes.io/projected/aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0-kube-api-access-wxz9d\") pod \"certified-operators-ggq4r\" (UID: \"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0\") " pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.464339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0-utilities\") pod \"certified-operators-ggq4r\" (UID: \"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0\") " pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.465398 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0-utilities\") pod \"certified-operators-ggq4r\" (UID: \"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0\") " pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.465641 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0-catalog-content\") pod \"certified-operators-ggq4r\" (UID: \"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0\") " pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.485619 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxz9d\" (UniqueName: \"kubernetes.io/projected/aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0-kube-api-access-wxz9d\") pod \"certified-operators-ggq4r\" (UID: \"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0\") " pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.591197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:24:55 crc kubenswrapper[4809]: W1127 17:24:55.990704 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaff3b1a0_ed26_4a23_b8fe_8ef38617c2d0.slice/crio-1fbc17306722d4c2e49014bb95a59a390910733e2ee82bd7ff579e0fe19f2e30 WatchSource:0}: Error finding container 1fbc17306722d4c2e49014bb95a59a390910733e2ee82bd7ff579e0fe19f2e30: Status 404 returned error can't find the container with id 1fbc17306722d4c2e49014bb95a59a390910733e2ee82bd7ff579e0fe19f2e30 Nov 27 17:24:55 crc kubenswrapper[4809]: I1127 17:24:55.994042 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ggq4r"] Nov 27 17:24:56 crc kubenswrapper[4809]: I1127 17:24:56.133128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggq4r" event={"ID":"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0","Type":"ContainerStarted","Data":"1fbc17306722d4c2e49014bb95a59a390910733e2ee82bd7ff579e0fe19f2e30"} Nov 27 17:24:56 crc kubenswrapper[4809]: I1127 17:24:56.143078 4809 generic.go:334] "Generic (PLEG): container finished" podID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerID="3cd6c5f92be49db875b69e57c16b2c93f852ee16738cacdef7c0f07f4d985ffc" exitCode=0 Nov 27 17:24:56 crc kubenswrapper[4809]: I1127 17:24:56.143134 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd5db" event={"ID":"7713e7ad-760d-4807-b2cf-17df91b637c0","Type":"ContainerDied","Data":"3cd6c5f92be49db875b69e57c16b2c93f852ee16738cacdef7c0f07f4d985ffc"} Nov 27 17:24:56 crc kubenswrapper[4809]: I1127 17:24:56.145622 4809 generic.go:334] "Generic (PLEG): container finished" podID="24a4f327-f4ee-45dd-938e-4de8c52b2e76" containerID="6eac39f8e74b1402518f74ff2e7a9409145d8b4e2b22ddc4becbd448055b3cdd" exitCode=0 Nov 27 17:24:56 crc kubenswrapper[4809]: I1127 17:24:56.145658 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk5rd" event={"ID":"24a4f327-f4ee-45dd-938e-4de8c52b2e76","Type":"ContainerDied","Data":"6eac39f8e74b1402518f74ff2e7a9409145d8b4e2b22ddc4becbd448055b3cdd"} Nov 27 17:24:58 crc kubenswrapper[4809]: I1127 17:24:58.161314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zfdv" event={"ID":"ac0a6dfd-901d-47b5-aa9e-44e040d13e75","Type":"ContainerStarted","Data":"73c8aa0af9b304b423cc618759ff2388f309d8915bc214b76d90690d6638b1f5"} Nov 27 17:24:58 crc kubenswrapper[4809]: I1127 17:24:58.162875 4809 generic.go:334] "Generic (PLEG): container finished" podID="aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0" containerID="b026eb09fef6dc246b685e082e41fe0d76ff258cd87e09e6d82cf9eec54be8bf" exitCode=0 Nov 27 17:24:58 crc kubenswrapper[4809]: I1127 17:24:58.162975 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggq4r" event={"ID":"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0","Type":"ContainerDied","Data":"b026eb09fef6dc246b685e082e41fe0d76ff258cd87e09e6d82cf9eec54be8bf"} Nov 27 17:24:58 crc kubenswrapper[4809]: I1127 17:24:58.166001 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8r4bc" event={"ID":"1b5febb2-32de-4c4a-a8fb-1168893feff9","Type":"ContainerStarted","Data":"42bca9782cd832cd7cd21608eabb4e2ac47bb86d5509d6b4b9d5f39e9fce0a74"} Nov 27 17:24:58 crc kubenswrapper[4809]: I1127 17:24:58.169096 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd5db" event={"ID":"7713e7ad-760d-4807-b2cf-17df91b637c0","Type":"ContainerStarted","Data":"07a6fccac267ea904673a762a7fe6a592cbdfe79ba5b734fd8f986fd179d7d9a"} Nov 27 17:24:58 crc kubenswrapper[4809]: I1127 17:24:58.172573 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hk5rd" event={"ID":"24a4f327-f4ee-45dd-938e-4de8c52b2e76","Type":"ContainerStarted","Data":"371f5a3516941decc4d1fd824f2bd1be4aa3f35b853813017bcab6ae2cb7632a"} Nov 27 17:24:58 crc kubenswrapper[4809]: I1127 17:24:58.227379 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hk5rd" podStartSLOduration=2.874352038 podStartE2EDuration="6.227351644s" podCreationTimestamp="2025-11-27 17:24:52 +0000 UTC" firstStartedPulling="2025-11-27 17:24:54.107797726 +0000 UTC m=+929.380255098" lastFinishedPulling="2025-11-27 17:24:57.460797352 +0000 UTC m=+932.733254704" observedRunningTime="2025-11-27 17:24:58.223417166 +0000 UTC m=+933.495874538" watchObservedRunningTime="2025-11-27 17:24:58.227351644 +0000 UTC m=+933.499808996" Nov 27 17:24:58 crc kubenswrapper[4809]: I1127 17:24:58.265924 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8r4bc" podStartSLOduration=3.081731123 podStartE2EDuration="7.265904087s" podCreationTimestamp="2025-11-27 17:24:51 +0000 UTC" firstStartedPulling="2025-11-27 17:24:53.093124913 +0000 UTC m=+928.365582265" lastFinishedPulling="2025-11-27 17:24:57.277297877 +0000 UTC m=+932.549755229" observedRunningTime="2025-11-27 17:24:58.246677733 +0000 UTC m=+933.519135085" watchObservedRunningTime="2025-11-27 17:24:58.265904087 +0000 UTC m=+933.538361439" Nov 27 17:24:59 crc kubenswrapper[4809]: I1127 17:24:59.185867 4809 generic.go:334] "Generic (PLEG): container finished" podID="ac0a6dfd-901d-47b5-aa9e-44e040d13e75" containerID="73c8aa0af9b304b423cc618759ff2388f309d8915bc214b76d90690d6638b1f5" exitCode=0 Nov 27 17:24:59 crc kubenswrapper[4809]: I1127 17:24:59.186791 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zfdv" event={"ID":"ac0a6dfd-901d-47b5-aa9e-44e040d13e75","Type":"ContainerDied","Data":"73c8aa0af9b304b423cc618759ff2388f309d8915bc214b76d90690d6638b1f5"} Nov 27 17:24:59 crc kubenswrapper[4809]: I1127 17:24:59.212534 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hd5db" podStartSLOduration=3.857874769 podStartE2EDuration="7.2125107s" podCreationTimestamp="2025-11-27 17:24:52 +0000 UTC" firstStartedPulling="2025-11-27 17:24:54.101424582 +0000 UTC m=+929.373881934" lastFinishedPulling="2025-11-27 17:24:57.456060513 +0000 UTC m=+932.728517865" observedRunningTime="2025-11-27 17:24:58.281915225 +0000 UTC m=+933.554372577" watchObservedRunningTime="2025-11-27 17:24:59.2125107 +0000 UTC m=+934.484968062" Nov 27 17:25:00 crc kubenswrapper[4809]: I1127 17:25:00.194109 4809 generic.go:334] "Generic (PLEG): container finished" podID="aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0" containerID="66d94bd369ca88edce547b2cce5fd5804efaccaa531eb2d1e7de6caad85894ef" exitCode=0 Nov 27 17:25:00 crc kubenswrapper[4809]: I1127 17:25:00.194174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggq4r" event={"ID":"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0","Type":"ContainerDied","Data":"66d94bd369ca88edce547b2cce5fd5804efaccaa531eb2d1e7de6caad85894ef"} Nov 27 17:25:01 crc kubenswrapper[4809]: I1127 17:25:01.345642 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:25:01 crc kubenswrapper[4809]: I1127 17:25:01.346011 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:25:01 crc kubenswrapper[4809]: I1127 17:25:01.385526 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:25:01 crc kubenswrapper[4809]: I1127 17:25:01.754772 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:25:01 crc kubenswrapper[4809]: I1127 17:25:01.755101 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:25:01 crc kubenswrapper[4809]: I1127 17:25:01.796682 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:25:02 crc kubenswrapper[4809]: I1127 17:25:02.217327 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zfdv" event={"ID":"ac0a6dfd-901d-47b5-aa9e-44e040d13e75","Type":"ContainerStarted","Data":"1b6d2ab8b3cafede0c02aa55a069780d5abdcad7a3dd56091b22ad38fbb6cea6"} Nov 27 17:25:02 crc kubenswrapper[4809]: I1127 17:25:02.222862 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggq4r" event={"ID":"aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0","Type":"ContainerStarted","Data":"e26fec511adfcda7d1011f8973d567620250cbcecc6a7bfe080c469bde19c669"} Nov 27 17:25:02 crc kubenswrapper[4809]: I1127 17:25:02.240088 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8zfdv" podStartSLOduration=2.9539341759999997 podStartE2EDuration="9.240068981s" podCreationTimestamp="2025-11-27 17:24:53 +0000 UTC" firstStartedPulling="2025-11-27 17:24:55.126814219 +0000 UTC m=+930.399271571" lastFinishedPulling="2025-11-27 17:25:01.412949024 +0000 UTC m=+936.685406376" observedRunningTime="2025-11-27 17:25:02.237290675 +0000 UTC m=+937.509748037" watchObservedRunningTime="2025-11-27 17:25:02.240068981 +0000 UTC m=+937.512526333" Nov 27 17:25:02 crc kubenswrapper[4809]: I1127 17:25:02.260782 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ggq4r" podStartSLOduration=3.867082398 podStartE2EDuration="7.260729375s" podCreationTimestamp="2025-11-27 17:24:55 +0000 UTC" firstStartedPulling="2025-11-27 17:24:58.164079515 +0000 UTC m=+933.436536867" lastFinishedPulling="2025-11-27 17:25:01.557726482 +0000 UTC m=+936.830183844" observedRunningTime="2025-11-27 17:25:02.258271539 +0000 UTC m=+937.530728891" watchObservedRunningTime="2025-11-27 17:25:02.260729375 +0000 UTC m=+937.533186727" Nov 27 17:25:02 crc kubenswrapper[4809]: I1127 17:25:02.264122 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-28jvg" Nov 27 17:25:02 crc kubenswrapper[4809]: I1127 17:25:02.269398 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:25:02 crc kubenswrapper[4809]: I1127 17:25:02.744533 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:25:02 crc kubenswrapper[4809]: I1127 17:25:02.745012 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:25:02 crc kubenswrapper[4809]: I1127 17:25:02.789276 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:25:03 crc kubenswrapper[4809]: I1127 17:25:03.156264 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:25:03 crc kubenswrapper[4809]: I1127 17:25:03.156652 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:25:03 crc kubenswrapper[4809]: I1127 17:25:03.202583 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:25:03 crc kubenswrapper[4809]: I1127 17:25:03.276221 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hk5rd" Nov 27 17:25:03 crc kubenswrapper[4809]: I1127 17:25:03.276512 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:25:04 crc kubenswrapper[4809]: I1127 17:25:04.140068 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:25:04 crc kubenswrapper[4809]: I1127 17:25:04.140115 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:25:04 crc kubenswrapper[4809]: I1127 17:25:04.216543 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8r4bc"] Nov 27 17:25:04 crc kubenswrapper[4809]: I1127 17:25:04.235628 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8r4bc" podUID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerName="registry-server" containerID="cri-o://42bca9782cd832cd7cd21608eabb4e2ac47bb86d5509d6b4b9d5f39e9fce0a74" gracePeriod=2 Nov 27 17:25:05 crc kubenswrapper[4809]: I1127 17:25:05.176587 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8zfdv" podUID="ac0a6dfd-901d-47b5-aa9e-44e040d13e75" containerName="registry-server" probeResult="failure" output=< Nov 27 17:25:05 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 27 17:25:05 crc kubenswrapper[4809]: > Nov 27 17:25:05 crc kubenswrapper[4809]: I1127 17:25:05.591632 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:25:05 crc kubenswrapper[4809]: I1127 17:25:05.592072 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:25:05 crc kubenswrapper[4809]: I1127 17:25:05.642707 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:25:06 crc kubenswrapper[4809]: I1127 17:25:06.282656 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ggq4r" Nov 27 17:25:06 crc kubenswrapper[4809]: I1127 17:25:06.615854 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd5db"] Nov 27 17:25:06 crc kubenswrapper[4809]: I1127 17:25:06.616104 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hd5db" podUID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerName="registry-server" containerID="cri-o://07a6fccac267ea904673a762a7fe6a592cbdfe79ba5b734fd8f986fd179d7d9a" gracePeriod=2 Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.259521 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8r4bc_1b5febb2-32de-4c4a-a8fb-1168893feff9/registry-server/0.log" Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.260654 4809 generic.go:334] "Generic (PLEG): container finished" podID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerID="42bca9782cd832cd7cd21608eabb4e2ac47bb86d5509d6b4b9d5f39e9fce0a74" exitCode=137 Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.260692 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8r4bc" event={"ID":"1b5febb2-32de-4c4a-a8fb-1168893feff9","Type":"ContainerDied","Data":"42bca9782cd832cd7cd21608eabb4e2ac47bb86d5509d6b4b9d5f39e9fce0a74"} Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.798335 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8r4bc_1b5febb2-32de-4c4a-a8fb-1168893feff9/registry-server/0.log" Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.799406 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.878578 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-catalog-content\") pod \"1b5febb2-32de-4c4a-a8fb-1168893feff9\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.878713 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4vgd\" (UniqueName: \"kubernetes.io/projected/1b5febb2-32de-4c4a-a8fb-1168893feff9-kube-api-access-k4vgd\") pod \"1b5febb2-32de-4c4a-a8fb-1168893feff9\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.878829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-utilities\") pod \"1b5febb2-32de-4c4a-a8fb-1168893feff9\" (UID: \"1b5febb2-32de-4c4a-a8fb-1168893feff9\") " Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.879684 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-utilities" (OuterVolumeSpecName: "utilities") pod "1b5febb2-32de-4c4a-a8fb-1168893feff9" (UID: "1b5febb2-32de-4c4a-a8fb-1168893feff9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.879987 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.897597 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b5febb2-32de-4c4a-a8fb-1168893feff9-kube-api-access-k4vgd" (OuterVolumeSpecName: "kube-api-access-k4vgd") pod "1b5febb2-32de-4c4a-a8fb-1168893feff9" (UID: "1b5febb2-32de-4c4a-a8fb-1168893feff9"). InnerVolumeSpecName "kube-api-access-k4vgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.933434 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b5febb2-32de-4c4a-a8fb-1168893feff9" (UID: "1b5febb2-32de-4c4a-a8fb-1168893feff9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.981733 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4vgd\" (UniqueName: \"kubernetes.io/projected/1b5febb2-32de-4c4a-a8fb-1168893feff9-kube-api-access-k4vgd\") on node \"crc\" DevicePath \"\"" Nov 27 17:25:08 crc kubenswrapper[4809]: I1127 17:25:08.981795 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5febb2-32de-4c4a-a8fb-1168893feff9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.279165 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8r4bc_1b5febb2-32de-4c4a-a8fb-1168893feff9/registry-server/0.log" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.280534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8r4bc" event={"ID":"1b5febb2-32de-4c4a-a8fb-1168893feff9","Type":"ContainerDied","Data":"c7722dba898a668fe4c982c933c9a44faab122f3501dcace610119ce4e39797d"} Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.280599 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8r4bc" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.280598 4809 scope.go:117] "RemoveContainer" containerID="42bca9782cd832cd7cd21608eabb4e2ac47bb86d5509d6b4b9d5f39e9fce0a74" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.283198 4809 generic.go:334] "Generic (PLEG): container finished" podID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerID="07a6fccac267ea904673a762a7fe6a592cbdfe79ba5b734fd8f986fd179d7d9a" exitCode=0 Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.283254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd5db" event={"ID":"7713e7ad-760d-4807-b2cf-17df91b637c0","Type":"ContainerDied","Data":"07a6fccac267ea904673a762a7fe6a592cbdfe79ba5b734fd8f986fd179d7d9a"} Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.312112 4809 scope.go:117] "RemoveContainer" containerID="47686c2ce8e17cd660aee93f35a87b1aba8fcddbbe86ce0cfd6c4f1758cbb684" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.356117 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8r4bc"] Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.358989 4809 scope.go:117] "RemoveContainer" containerID="aa22dcf997faf7d8f05c1279b6e9c13a17615b401f95b58679d418ad719cccb5" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.367907 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8r4bc"] Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.467634 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b5febb2-32de-4c4a-a8fb-1168893feff9" path="/var/lib/kubelet/pods/1b5febb2-32de-4c4a-a8fb-1168893feff9/volumes" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.706998 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.792325 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5jv2\" (UniqueName: \"kubernetes.io/projected/7713e7ad-760d-4807-b2cf-17df91b637c0-kube-api-access-t5jv2\") pod \"7713e7ad-760d-4807-b2cf-17df91b637c0\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.792425 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-catalog-content\") pod \"7713e7ad-760d-4807-b2cf-17df91b637c0\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.792517 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-utilities\") pod \"7713e7ad-760d-4807-b2cf-17df91b637c0\" (UID: \"7713e7ad-760d-4807-b2cf-17df91b637c0\") " Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.793468 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-utilities" (OuterVolumeSpecName: "utilities") pod "7713e7ad-760d-4807-b2cf-17df91b637c0" (UID: "7713e7ad-760d-4807-b2cf-17df91b637c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.796884 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7713e7ad-760d-4807-b2cf-17df91b637c0-kube-api-access-t5jv2" (OuterVolumeSpecName: "kube-api-access-t5jv2") pod "7713e7ad-760d-4807-b2cf-17df91b637c0" (UID: "7713e7ad-760d-4807-b2cf-17df91b637c0"). InnerVolumeSpecName "kube-api-access-t5jv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.811441 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7713e7ad-760d-4807-b2cf-17df91b637c0" (UID: "7713e7ad-760d-4807-b2cf-17df91b637c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.893911 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.893961 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5jv2\" (UniqueName: \"kubernetes.io/projected/7713e7ad-760d-4807-b2cf-17df91b637c0-kube-api-access-t5jv2\") on node \"crc\" DevicePath \"\"" Nov 27 17:25:09 crc kubenswrapper[4809]: I1127 17:25:09.893972 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7713e7ad-760d-4807-b2cf-17df91b637c0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:25:10 crc kubenswrapper[4809]: I1127 17:25:10.295227 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hd5db" event={"ID":"7713e7ad-760d-4807-b2cf-17df91b637c0","Type":"ContainerDied","Data":"52d10367defdf2ca54d855baddb743a6d7031b8079bc3020f346ec2b5a3d6b7d"} Nov 27 17:25:10 crc kubenswrapper[4809]: I1127 17:25:10.295300 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hd5db" Nov 27 17:25:10 crc kubenswrapper[4809]: I1127 17:25:10.295338 4809 scope.go:117] "RemoveContainer" containerID="07a6fccac267ea904673a762a7fe6a592cbdfe79ba5b734fd8f986fd179d7d9a" Nov 27 17:25:10 crc kubenswrapper[4809]: I1127 17:25:10.311819 4809 scope.go:117] "RemoveContainer" containerID="3cd6c5f92be49db875b69e57c16b2c93f852ee16738cacdef7c0f07f4d985ffc" Nov 27 17:25:10 crc kubenswrapper[4809]: I1127 17:25:10.329272 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd5db"] Nov 27 17:25:10 crc kubenswrapper[4809]: I1127 17:25:10.333590 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hd5db"] Nov 27 17:25:10 crc kubenswrapper[4809]: I1127 17:25:10.342502 4809 scope.go:117] "RemoveContainer" containerID="c0193f4086a576d33312924f176de8c41d1706b5d0e03cfa7bd9efa79794c915" Nov 27 17:25:11 crc kubenswrapper[4809]: I1127 17:25:11.466565 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7713e7ad-760d-4807-b2cf-17df91b637c0" path="/var/lib/kubelet/pods/7713e7ad-760d-4807-b2cf-17df91b637c0/volumes" Nov 27 17:25:14 crc kubenswrapper[4809]: I1127 17:25:14.184909 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:25:14 crc kubenswrapper[4809]: I1127 17:25:14.226558 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8zfdv" Nov 27 17:25:55 crc kubenswrapper[4809]: I1127 17:25:55.779282 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:25:55 crc kubenswrapper[4809]: I1127 17:25:55.779695 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:26:25 crc kubenswrapper[4809]: I1127 17:26:25.779449 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:26:25 crc kubenswrapper[4809]: I1127 17:26:25.780066 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:26:55 crc kubenswrapper[4809]: I1127 17:26:55.779857 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:26:55 crc kubenswrapper[4809]: I1127 17:26:55.780458 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:26:55 crc kubenswrapper[4809]: I1127 17:26:55.780501 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:26:55 crc kubenswrapper[4809]: I1127 17:26:55.780983 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f6a9abdc5c8929f8e6588fb217c1a8bad8386fe8d57f874960d924240fd36c12"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:26:55 crc kubenswrapper[4809]: I1127 17:26:55.781033 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://f6a9abdc5c8929f8e6588fb217c1a8bad8386fe8d57f874960d924240fd36c12" gracePeriod=600 Nov 27 17:26:56 crc kubenswrapper[4809]: I1127 17:26:56.373093 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="f6a9abdc5c8929f8e6588fb217c1a8bad8386fe8d57f874960d924240fd36c12" exitCode=0 Nov 27 17:26:56 crc kubenswrapper[4809]: I1127 17:26:56.373496 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"f6a9abdc5c8929f8e6588fb217c1a8bad8386fe8d57f874960d924240fd36c12"} Nov 27 17:26:56 crc kubenswrapper[4809]: I1127 17:26:56.373563 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"0f129562b4adcd0d1caac594ce6928b2856438351f75afb41cbf7f3b99003516"} Nov 27 17:26:56 crc kubenswrapper[4809]: I1127 17:26:56.373581 4809 scope.go:117] "RemoveContainer" containerID="4c5bbc3124a8972c9912c532472d23e13e06555c5f6e029a3731193652f77944" Nov 27 17:28:25 crc kubenswrapper[4809]: I1127 17:28:25.861961 4809 scope.go:117] "RemoveContainer" containerID="75614c86950abf0ee4617396011e85116532be6765626db5c2d9e220b2cf5681" Nov 27 17:28:25 crc kubenswrapper[4809]: I1127 17:28:25.890420 4809 scope.go:117] "RemoveContainer" containerID="5be04991360112cd9f3739ab2219da8058377222e3ecfe5dc987797c1495cb5c" Nov 27 17:28:25 crc kubenswrapper[4809]: I1127 17:28:25.911763 4809 scope.go:117] "RemoveContainer" containerID="c685e71e36534e3ee49bd908df32f0a7be883fd0809d55cee320cfecee1ae6fe" Nov 27 17:28:25 crc kubenswrapper[4809]: I1127 17:28:25.933908 4809 scope.go:117] "RemoveContainer" containerID="63971d2b1732f89c36af664e2ee23b5b2331e4d2d559ae61c3cd6180d32ab6b4" Nov 27 17:28:25 crc kubenswrapper[4809]: I1127 17:28:25.955757 4809 scope.go:117] "RemoveContainer" containerID="9bc85c58f44901adac8e32062f4798528a77e3e5c866710514bb560e4922f5d4" Nov 27 17:28:25 crc kubenswrapper[4809]: I1127 17:28:25.978919 4809 scope.go:117] "RemoveContainer" containerID="12b8d7f66e8ede38a941826bb3e002ea2c0b2b2b8f3783bd3ae6fd25cbe87795" Nov 27 17:29:25 crc kubenswrapper[4809]: I1127 17:29:25.780200 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:29:25 crc kubenswrapper[4809]: I1127 17:29:25.781065 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:29:55 crc kubenswrapper[4809]: I1127 17:29:55.779254 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:29:55 crc kubenswrapper[4809]: I1127 17:29:55.779726 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.141783 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp"] Nov 27 17:30:00 crc kubenswrapper[4809]: E1127 17:30:00.142377 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerName="extract-utilities" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.142393 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerName="extract-utilities" Nov 27 17:30:00 crc kubenswrapper[4809]: E1127 17:30:00.142404 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerName="extract-content" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.142412 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerName="extract-content" Nov 27 17:30:00 crc kubenswrapper[4809]: E1127 17:30:00.142426 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerName="registry-server" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.142434 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerName="registry-server" Nov 27 17:30:00 crc kubenswrapper[4809]: E1127 17:30:00.142446 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerName="extract-content" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.142493 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerName="extract-content" Nov 27 17:30:00 crc kubenswrapper[4809]: E1127 17:30:00.142502 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerName="extract-utilities" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.142508 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerName="extract-utilities" Nov 27 17:30:00 crc kubenswrapper[4809]: E1127 17:30:00.142517 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerName="registry-server" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.142522 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerName="registry-server" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.142618 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b5febb2-32de-4c4a-a8fb-1168893feff9" containerName="registry-server" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.142631 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7713e7ad-760d-4807-b2cf-17df91b637c0" containerName="registry-server" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.143093 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.145063 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.145965 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.150110 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp"] Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.274189 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb1409d8-a556-486c-8a94-8e5caf14f196-secret-volume\") pod \"collect-profiles-29404410-s7mpp\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.274289 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wkcl\" (UniqueName: \"kubernetes.io/projected/bb1409d8-a556-486c-8a94-8e5caf14f196-kube-api-access-4wkcl\") pod \"collect-profiles-29404410-s7mpp\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.274337 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb1409d8-a556-486c-8a94-8e5caf14f196-config-volume\") pod \"collect-profiles-29404410-s7mpp\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.375927 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb1409d8-a556-486c-8a94-8e5caf14f196-secret-volume\") pod \"collect-profiles-29404410-s7mpp\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.376005 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wkcl\" (UniqueName: \"kubernetes.io/projected/bb1409d8-a556-486c-8a94-8e5caf14f196-kube-api-access-4wkcl\") pod \"collect-profiles-29404410-s7mpp\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.376042 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb1409d8-a556-486c-8a94-8e5caf14f196-config-volume\") pod \"collect-profiles-29404410-s7mpp\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.376918 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb1409d8-a556-486c-8a94-8e5caf14f196-config-volume\") pod \"collect-profiles-29404410-s7mpp\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.382571 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb1409d8-a556-486c-8a94-8e5caf14f196-secret-volume\") pod \"collect-profiles-29404410-s7mpp\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.391109 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wkcl\" (UniqueName: \"kubernetes.io/projected/bb1409d8-a556-486c-8a94-8e5caf14f196-kube-api-access-4wkcl\") pod \"collect-profiles-29404410-s7mpp\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.464408 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:00 crc kubenswrapper[4809]: I1127 17:30:00.872056 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp"] Nov 27 17:30:01 crc kubenswrapper[4809]: I1127 17:30:01.425545 4809 generic.go:334] "Generic (PLEG): container finished" podID="bb1409d8-a556-486c-8a94-8e5caf14f196" containerID="c779647fad2e5a6faeed63d86d563effd014d8ffe393623c81d00fe6a316d4c1" exitCode=0 Nov 27 17:30:01 crc kubenswrapper[4809]: I1127 17:30:01.425656 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" event={"ID":"bb1409d8-a556-486c-8a94-8e5caf14f196","Type":"ContainerDied","Data":"c779647fad2e5a6faeed63d86d563effd014d8ffe393623c81d00fe6a316d4c1"} Nov 27 17:30:01 crc kubenswrapper[4809]: I1127 17:30:01.426052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" event={"ID":"bb1409d8-a556-486c-8a94-8e5caf14f196","Type":"ContainerStarted","Data":"c29833321fe4a424a280f37a609c853b7b899ad7bb88b42054157f9db6ffc83b"} Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.728023 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.815138 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wkcl\" (UniqueName: \"kubernetes.io/projected/bb1409d8-a556-486c-8a94-8e5caf14f196-kube-api-access-4wkcl\") pod \"bb1409d8-a556-486c-8a94-8e5caf14f196\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.815609 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb1409d8-a556-486c-8a94-8e5caf14f196-config-volume\") pod \"bb1409d8-a556-486c-8a94-8e5caf14f196\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.815734 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb1409d8-a556-486c-8a94-8e5caf14f196-secret-volume\") pod \"bb1409d8-a556-486c-8a94-8e5caf14f196\" (UID: \"bb1409d8-a556-486c-8a94-8e5caf14f196\") " Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.816624 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb1409d8-a556-486c-8a94-8e5caf14f196-config-volume" (OuterVolumeSpecName: "config-volume") pod "bb1409d8-a556-486c-8a94-8e5caf14f196" (UID: "bb1409d8-a556-486c-8a94-8e5caf14f196"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.824367 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb1409d8-a556-486c-8a94-8e5caf14f196-kube-api-access-4wkcl" (OuterVolumeSpecName: "kube-api-access-4wkcl") pod "bb1409d8-a556-486c-8a94-8e5caf14f196" (UID: "bb1409d8-a556-486c-8a94-8e5caf14f196"). InnerVolumeSpecName "kube-api-access-4wkcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.824893 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb1409d8-a556-486c-8a94-8e5caf14f196-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bb1409d8-a556-486c-8a94-8e5caf14f196" (UID: "bb1409d8-a556-486c-8a94-8e5caf14f196"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.917475 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb1409d8-a556-486c-8a94-8e5caf14f196-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.917524 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wkcl\" (UniqueName: \"kubernetes.io/projected/bb1409d8-a556-486c-8a94-8e5caf14f196-kube-api-access-4wkcl\") on node \"crc\" DevicePath \"\"" Nov 27 17:30:02 crc kubenswrapper[4809]: I1127 17:30:02.917536 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb1409d8-a556-486c-8a94-8e5caf14f196-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 17:30:03 crc kubenswrapper[4809]: I1127 17:30:03.440869 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" event={"ID":"bb1409d8-a556-486c-8a94-8e5caf14f196","Type":"ContainerDied","Data":"c29833321fe4a424a280f37a609c853b7b899ad7bb88b42054157f9db6ffc83b"} Nov 27 17:30:03 crc kubenswrapper[4809]: I1127 17:30:03.440914 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c29833321fe4a424a280f37a609c853b7b899ad7bb88b42054157f9db6ffc83b" Nov 27 17:30:03 crc kubenswrapper[4809]: I1127 17:30:03.440942 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp" Nov 27 17:30:25 crc kubenswrapper[4809]: I1127 17:30:25.779484 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:30:25 crc kubenswrapper[4809]: I1127 17:30:25.780085 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:30:25 crc kubenswrapper[4809]: I1127 17:30:25.780131 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:30:25 crc kubenswrapper[4809]: I1127 17:30:25.780811 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f129562b4adcd0d1caac594ce6928b2856438351f75afb41cbf7f3b99003516"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:30:25 crc kubenswrapper[4809]: I1127 17:30:25.780863 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://0f129562b4adcd0d1caac594ce6928b2856438351f75afb41cbf7f3b99003516" gracePeriod=600 Nov 27 17:30:26 crc kubenswrapper[4809]: I1127 17:30:26.589725 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="0f129562b4adcd0d1caac594ce6928b2856438351f75afb41cbf7f3b99003516" exitCode=0 Nov 27 17:30:26 crc kubenswrapper[4809]: I1127 17:30:26.589791 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"0f129562b4adcd0d1caac594ce6928b2856438351f75afb41cbf7f3b99003516"} Nov 27 17:30:26 crc kubenswrapper[4809]: I1127 17:30:26.590180 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"de4c3ebc223ee9f9580cd86c386eb6d52d4734548883eb67a85381672ea09fa4"} Nov 27 17:30:26 crc kubenswrapper[4809]: I1127 17:30:26.590215 4809 scope.go:117] "RemoveContainer" containerID="f6a9abdc5c8929f8e6588fb217c1a8bad8386fe8d57f874960d924240fd36c12" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.414845 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-48p8j"] Nov 27 17:32:23 crc kubenswrapper[4809]: E1127 17:32:23.415548 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1409d8-a556-486c-8a94-8e5caf14f196" containerName="collect-profiles" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.415562 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1409d8-a556-486c-8a94-8e5caf14f196" containerName="collect-profiles" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.415707 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb1409d8-a556-486c-8a94-8e5caf14f196" containerName="collect-profiles" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.416676 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.452989 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48p8j"] Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.530252 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-utilities\") pod \"redhat-operators-48p8j\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.530354 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-catalog-content\") pod \"redhat-operators-48p8j\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.530453 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkkkr\" (UniqueName: \"kubernetes.io/projected/850f5b12-ec46-441d-ae20-de371b76f93b-kube-api-access-fkkkr\") pod \"redhat-operators-48p8j\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.631365 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkkkr\" (UniqueName: \"kubernetes.io/projected/850f5b12-ec46-441d-ae20-de371b76f93b-kube-api-access-fkkkr\") pod \"redhat-operators-48p8j\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.631429 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-utilities\") pod \"redhat-operators-48p8j\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.631465 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-catalog-content\") pod \"redhat-operators-48p8j\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.631974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-catalog-content\") pod \"redhat-operators-48p8j\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.632390 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-utilities\") pod \"redhat-operators-48p8j\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.655592 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkkkr\" (UniqueName: \"kubernetes.io/projected/850f5b12-ec46-441d-ae20-de371b76f93b-kube-api-access-fkkkr\") pod \"redhat-operators-48p8j\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.739670 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:23 crc kubenswrapper[4809]: I1127 17:32:23.943037 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48p8j"] Nov 27 17:32:24 crc kubenswrapper[4809]: I1127 17:32:24.286554 4809 generic.go:334] "Generic (PLEG): container finished" podID="850f5b12-ec46-441d-ae20-de371b76f93b" containerID="4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362" exitCode=0 Nov 27 17:32:24 crc kubenswrapper[4809]: I1127 17:32:24.286593 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48p8j" event={"ID":"850f5b12-ec46-441d-ae20-de371b76f93b","Type":"ContainerDied","Data":"4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362"} Nov 27 17:32:24 crc kubenswrapper[4809]: I1127 17:32:24.286616 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48p8j" event={"ID":"850f5b12-ec46-441d-ae20-de371b76f93b","Type":"ContainerStarted","Data":"f5d3aedbe37dfad98f76b9261777cb3c7d7d495d85adec93d2d18f436fb94dc6"} Nov 27 17:32:24 crc kubenswrapper[4809]: I1127 17:32:24.288577 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 17:32:25 crc kubenswrapper[4809]: I1127 17:32:25.293579 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48p8j" event={"ID":"850f5b12-ec46-441d-ae20-de371b76f93b","Type":"ContainerStarted","Data":"d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d"} Nov 27 17:32:26 crc kubenswrapper[4809]: I1127 17:32:26.303502 4809 generic.go:334] "Generic (PLEG): container finished" podID="850f5b12-ec46-441d-ae20-de371b76f93b" containerID="d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d" exitCode=0 Nov 27 17:32:26 crc kubenswrapper[4809]: I1127 17:32:26.303543 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48p8j" event={"ID":"850f5b12-ec46-441d-ae20-de371b76f93b","Type":"ContainerDied","Data":"d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d"} Nov 27 17:32:27 crc kubenswrapper[4809]: I1127 17:32:27.312028 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48p8j" event={"ID":"850f5b12-ec46-441d-ae20-de371b76f93b","Type":"ContainerStarted","Data":"9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf"} Nov 27 17:32:27 crc kubenswrapper[4809]: I1127 17:32:27.337133 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-48p8j" podStartSLOduration=1.77281239 podStartE2EDuration="4.337111754s" podCreationTimestamp="2025-11-27 17:32:23 +0000 UTC" firstStartedPulling="2025-11-27 17:32:24.288378994 +0000 UTC m=+1379.560836346" lastFinishedPulling="2025-11-27 17:32:26.852678368 +0000 UTC m=+1382.125135710" observedRunningTime="2025-11-27 17:32:27.333533326 +0000 UTC m=+1382.605990688" watchObservedRunningTime="2025-11-27 17:32:27.337111754 +0000 UTC m=+1382.609569116" Nov 27 17:32:33 crc kubenswrapper[4809]: I1127 17:32:33.740247 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:33 crc kubenswrapper[4809]: I1127 17:32:33.740894 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:33 crc kubenswrapper[4809]: I1127 17:32:33.778115 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:34 crc kubenswrapper[4809]: I1127 17:32:34.398811 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:34 crc kubenswrapper[4809]: I1127 17:32:34.438786 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48p8j"] Nov 27 17:32:36 crc kubenswrapper[4809]: I1127 17:32:36.361455 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-48p8j" podUID="850f5b12-ec46-441d-ae20-de371b76f93b" containerName="registry-server" containerID="cri-o://9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf" gracePeriod=2 Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.205611 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.336232 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-catalog-content\") pod \"850f5b12-ec46-441d-ae20-de371b76f93b\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.337099 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-utilities\") pod \"850f5b12-ec46-441d-ae20-de371b76f93b\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.337227 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkkkr\" (UniqueName: \"kubernetes.io/projected/850f5b12-ec46-441d-ae20-de371b76f93b-kube-api-access-fkkkr\") pod \"850f5b12-ec46-441d-ae20-de371b76f93b\" (UID: \"850f5b12-ec46-441d-ae20-de371b76f93b\") " Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.338827 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-utilities" (OuterVolumeSpecName: "utilities") pod "850f5b12-ec46-441d-ae20-de371b76f93b" (UID: "850f5b12-ec46-441d-ae20-de371b76f93b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.343665 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/850f5b12-ec46-441d-ae20-de371b76f93b-kube-api-access-fkkkr" (OuterVolumeSpecName: "kube-api-access-fkkkr") pod "850f5b12-ec46-441d-ae20-de371b76f93b" (UID: "850f5b12-ec46-441d-ae20-de371b76f93b"). InnerVolumeSpecName "kube-api-access-fkkkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.370107 4809 generic.go:334] "Generic (PLEG): container finished" podID="850f5b12-ec46-441d-ae20-de371b76f93b" containerID="9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf" exitCode=0 Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.370153 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48p8j" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.370164 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48p8j" event={"ID":"850f5b12-ec46-441d-ae20-de371b76f93b","Type":"ContainerDied","Data":"9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf"} Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.371004 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48p8j" event={"ID":"850f5b12-ec46-441d-ae20-de371b76f93b","Type":"ContainerDied","Data":"f5d3aedbe37dfad98f76b9261777cb3c7d7d495d85adec93d2d18f436fb94dc6"} Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.371027 4809 scope.go:117] "RemoveContainer" containerID="9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.401549 4809 scope.go:117] "RemoveContainer" containerID="d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.420847 4809 scope.go:117] "RemoveContainer" containerID="4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.439301 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkkkr\" (UniqueName: \"kubernetes.io/projected/850f5b12-ec46-441d-ae20-de371b76f93b-kube-api-access-fkkkr\") on node \"crc\" DevicePath \"\"" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.439329 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.439387 4809 scope.go:117] "RemoveContainer" containerID="9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf" Nov 27 17:32:37 crc kubenswrapper[4809]: E1127 17:32:37.439952 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf\": container with ID starting with 9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf not found: ID does not exist" containerID="9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.439990 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf"} err="failed to get container status \"9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf\": rpc error: code = NotFound desc = could not find container \"9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf\": container with ID starting with 9c07df7467421a3fe4d644d6044772417fa820d5d99a923684956b1abf190acf not found: ID does not exist" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.440019 4809 scope.go:117] "RemoveContainer" containerID="d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d" Nov 27 17:32:37 crc kubenswrapper[4809]: E1127 17:32:37.440354 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d\": container with ID starting with d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d not found: ID does not exist" containerID="d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.440411 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d"} err="failed to get container status \"d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d\": rpc error: code = NotFound desc = could not find container \"d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d\": container with ID starting with d277717ec04e8f9bc7e3567f3b3191e86d0fa92775cef682c92fc264cdf9fd1d not found: ID does not exist" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.440439 4809 scope.go:117] "RemoveContainer" containerID="4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362" Nov 27 17:32:37 crc kubenswrapper[4809]: E1127 17:32:37.440715 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362\": container with ID starting with 4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362 not found: ID does not exist" containerID="4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.440780 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362"} err="failed to get container status \"4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362\": rpc error: code = NotFound desc = could not find container \"4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362\": container with ID starting with 4c146a7f84425e71889e9dd5ecbbb9d575d26319afec0d78aa5263566f1d2362 not found: ID does not exist" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.454997 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "850f5b12-ec46-441d-ae20-de371b76f93b" (UID: "850f5b12-ec46-441d-ae20-de371b76f93b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.540471 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850f5b12-ec46-441d-ae20-de371b76f93b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.692928 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48p8j"] Nov 27 17:32:37 crc kubenswrapper[4809]: I1127 17:32:37.699649 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-48p8j"] Nov 27 17:32:39 crc kubenswrapper[4809]: I1127 17:32:39.468337 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="850f5b12-ec46-441d-ae20-de371b76f93b" path="/var/lib/kubelet/pods/850f5b12-ec46-441d-ae20-de371b76f93b/volumes" Nov 27 17:32:55 crc kubenswrapper[4809]: I1127 17:32:55.779303 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:32:55 crc kubenswrapper[4809]: I1127 17:32:55.779862 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.379425 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj"] Nov 27 17:33:13 crc kubenswrapper[4809]: E1127 17:33:13.381150 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850f5b12-ec46-441d-ae20-de371b76f93b" containerName="extract-utilities" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.381224 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="850f5b12-ec46-441d-ae20-de371b76f93b" containerName="extract-utilities" Nov 27 17:33:13 crc kubenswrapper[4809]: E1127 17:33:13.381295 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850f5b12-ec46-441d-ae20-de371b76f93b" containerName="extract-content" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.381351 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="850f5b12-ec46-441d-ae20-de371b76f93b" containerName="extract-content" Nov 27 17:33:13 crc kubenswrapper[4809]: E1127 17:33:13.381420 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850f5b12-ec46-441d-ae20-de371b76f93b" containerName="registry-server" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.381480 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="850f5b12-ec46-441d-ae20-de371b76f93b" containerName="registry-server" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.381626 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="850f5b12-ec46-441d-ae20-de371b76f93b" containerName="registry-server" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.382540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.385019 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.394110 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj"] Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.470037 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.470101 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.470161 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p7mn\" (UniqueName: \"kubernetes.io/projected/96b8cee9-c201-410d-9391-cc02bb9a2eeb-kube-api-access-9p7mn\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.571613 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p7mn\" (UniqueName: \"kubernetes.io/projected/96b8cee9-c201-410d-9391-cc02bb9a2eeb-kube-api-access-9p7mn\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.571768 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.571808 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.572382 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.572677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.599969 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p7mn\" (UniqueName: \"kubernetes.io/projected/96b8cee9-c201-410d-9391-cc02bb9a2eeb-kube-api-access-9p7mn\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.704850 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:13 crc kubenswrapper[4809]: I1127 17:33:13.900372 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj"] Nov 27 17:33:14 crc kubenswrapper[4809]: I1127 17:33:14.619172 4809 generic.go:334] "Generic (PLEG): container finished" podID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerID="f556bcd8d6d3b6c8a978ecd914c88eb8df003001520b946f88c7f24fbf1a9f41" exitCode=0 Nov 27 17:33:14 crc kubenswrapper[4809]: I1127 17:33:14.619215 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" event={"ID":"96b8cee9-c201-410d-9391-cc02bb9a2eeb","Type":"ContainerDied","Data":"f556bcd8d6d3b6c8a978ecd914c88eb8df003001520b946f88c7f24fbf1a9f41"} Nov 27 17:33:14 crc kubenswrapper[4809]: I1127 17:33:14.619240 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" event={"ID":"96b8cee9-c201-410d-9391-cc02bb9a2eeb","Type":"ContainerStarted","Data":"ce8e8352fbe222ac9fc37cf07ebea42c222757ee38ccc72e54e674207d592708"} Nov 27 17:33:16 crc kubenswrapper[4809]: I1127 17:33:16.635591 4809 generic.go:334] "Generic (PLEG): container finished" podID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerID="13fe0115a756bd9c40a15216c2a3d969d852a4edf66dbcc8061ba646e96f9a3a" exitCode=0 Nov 27 17:33:16 crc kubenswrapper[4809]: I1127 17:33:16.635690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" event={"ID":"96b8cee9-c201-410d-9391-cc02bb9a2eeb","Type":"ContainerDied","Data":"13fe0115a756bd9c40a15216c2a3d969d852a4edf66dbcc8061ba646e96f9a3a"} Nov 27 17:33:17 crc kubenswrapper[4809]: I1127 17:33:17.645532 4809 generic.go:334] "Generic (PLEG): container finished" podID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerID="2067b0423737dc9003371527866539f38a3c08e9148d21d89b8ae90fcea69524" exitCode=0 Nov 27 17:33:17 crc kubenswrapper[4809]: I1127 17:33:17.645578 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" event={"ID":"96b8cee9-c201-410d-9391-cc02bb9a2eeb","Type":"ContainerDied","Data":"2067b0423737dc9003371527866539f38a3c08e9148d21d89b8ae90fcea69524"} Nov 27 17:33:18 crc kubenswrapper[4809]: I1127 17:33:18.899879 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:18 crc kubenswrapper[4809]: I1127 17:33:18.948792 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-bundle\") pod \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " Nov 27 17:33:18 crc kubenswrapper[4809]: I1127 17:33:18.948919 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-util\") pod \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " Nov 27 17:33:18 crc kubenswrapper[4809]: I1127 17:33:18.948987 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p7mn\" (UniqueName: \"kubernetes.io/projected/96b8cee9-c201-410d-9391-cc02bb9a2eeb-kube-api-access-9p7mn\") pod \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\" (UID: \"96b8cee9-c201-410d-9391-cc02bb9a2eeb\") " Nov 27 17:33:18 crc kubenswrapper[4809]: I1127 17:33:18.949538 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-bundle" (OuterVolumeSpecName: "bundle") pod "96b8cee9-c201-410d-9391-cc02bb9a2eeb" (UID: "96b8cee9-c201-410d-9391-cc02bb9a2eeb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:33:18 crc kubenswrapper[4809]: I1127 17:33:18.956892 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b8cee9-c201-410d-9391-cc02bb9a2eeb-kube-api-access-9p7mn" (OuterVolumeSpecName: "kube-api-access-9p7mn") pod "96b8cee9-c201-410d-9391-cc02bb9a2eeb" (UID: "96b8cee9-c201-410d-9391-cc02bb9a2eeb"). InnerVolumeSpecName "kube-api-access-9p7mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:33:18 crc kubenswrapper[4809]: I1127 17:33:18.982351 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-util" (OuterVolumeSpecName: "util") pod "96b8cee9-c201-410d-9391-cc02bb9a2eeb" (UID: "96b8cee9-c201-410d-9391-cc02bb9a2eeb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:33:19 crc kubenswrapper[4809]: I1127 17:33:19.050947 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p7mn\" (UniqueName: \"kubernetes.io/projected/96b8cee9-c201-410d-9391-cc02bb9a2eeb-kube-api-access-9p7mn\") on node \"crc\" DevicePath \"\"" Nov 27 17:33:19 crc kubenswrapper[4809]: I1127 17:33:19.050995 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:33:19 crc kubenswrapper[4809]: I1127 17:33:19.051006 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96b8cee9-c201-410d-9391-cc02bb9a2eeb-util\") on node \"crc\" DevicePath \"\"" Nov 27 17:33:19 crc kubenswrapper[4809]: I1127 17:33:19.659681 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" event={"ID":"96b8cee9-c201-410d-9391-cc02bb9a2eeb","Type":"ContainerDied","Data":"ce8e8352fbe222ac9fc37cf07ebea42c222757ee38ccc72e54e674207d592708"} Nov 27 17:33:19 crc kubenswrapper[4809]: I1127 17:33:19.659725 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce8e8352fbe222ac9fc37cf07ebea42c222757ee38ccc72e54e674207d592708" Nov 27 17:33:19 crc kubenswrapper[4809]: I1127 17:33:19.659840 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.821078 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf"] Nov 27 17:33:21 crc kubenswrapper[4809]: E1127 17:33:21.821564 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerName="util" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.821575 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerName="util" Nov 27 17:33:21 crc kubenswrapper[4809]: E1127 17:33:21.821587 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerName="extract" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.821593 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerName="extract" Nov 27 17:33:21 crc kubenswrapper[4809]: E1127 17:33:21.821613 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerName="pull" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.821620 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerName="pull" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.821724 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="96b8cee9-c201-410d-9391-cc02bb9a2eeb" containerName="extract" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.822174 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.824354 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.825470 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.825524 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-dkg67" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.834756 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf"] Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.890091 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z76bn\" (UniqueName: \"kubernetes.io/projected/aca17653-5316-4c78-8062-9526abf28d97-kube-api-access-z76bn\") pod \"nmstate-operator-5b5b58f5c8-pz6jf\" (UID: \"aca17653-5316-4c78-8062-9526abf28d97\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf" Nov 27 17:33:21 crc kubenswrapper[4809]: I1127 17:33:21.991492 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z76bn\" (UniqueName: \"kubernetes.io/projected/aca17653-5316-4c78-8062-9526abf28d97-kube-api-access-z76bn\") pod \"nmstate-operator-5b5b58f5c8-pz6jf\" (UID: \"aca17653-5316-4c78-8062-9526abf28d97\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf" Nov 27 17:33:22 crc kubenswrapper[4809]: I1127 17:33:22.009871 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z76bn\" (UniqueName: \"kubernetes.io/projected/aca17653-5316-4c78-8062-9526abf28d97-kube-api-access-z76bn\") pod \"nmstate-operator-5b5b58f5c8-pz6jf\" (UID: \"aca17653-5316-4c78-8062-9526abf28d97\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf" Nov 27 17:33:22 crc kubenswrapper[4809]: I1127 17:33:22.136603 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf" Nov 27 17:33:22 crc kubenswrapper[4809]: I1127 17:33:22.796047 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf"] Nov 27 17:33:23 crc kubenswrapper[4809]: I1127 17:33:23.683442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf" event={"ID":"aca17653-5316-4c78-8062-9526abf28d97","Type":"ContainerStarted","Data":"bf85df42a132eb97acae3cd9b9d1fa7c41f456641a7e4d1af19cc44af04d2f19"} Nov 27 17:33:25 crc kubenswrapper[4809]: I1127 17:33:25.779938 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:33:25 crc kubenswrapper[4809]: I1127 17:33:25.780503 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:33:26 crc kubenswrapper[4809]: I1127 17:33:26.717564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf" event={"ID":"aca17653-5316-4c78-8062-9526abf28d97","Type":"ContainerStarted","Data":"a187549e31c51af576a873883a3849e99738e2d33ae876df36f83719bb4da07e"} Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.662311 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-pz6jf" podStartSLOduration=3.8814129680000002 podStartE2EDuration="6.662287487s" podCreationTimestamp="2025-11-27 17:33:21 +0000 UTC" firstStartedPulling="2025-11-27 17:33:22.814731807 +0000 UTC m=+1438.087189159" lastFinishedPulling="2025-11-27 17:33:25.595606326 +0000 UTC m=+1440.868063678" observedRunningTime="2025-11-27 17:33:26.732408158 +0000 UTC m=+1442.004865530" watchObservedRunningTime="2025-11-27 17:33:27.662287487 +0000 UTC m=+1442.934744839" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.664615 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm"] Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.666051 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.671702 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-lb5hq" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.675007 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm"] Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.680966 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5"] Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.682027 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.683436 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.698932 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-sccnl"] Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.700328 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.704075 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5"] Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.779665 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf924\" (UniqueName: \"kubernetes.io/projected/637b5bb3-b183-43dd-bca4-0349ac11e7d5-kube-api-access-vf924\") pod \"nmstate-metrics-7f946cbc9-52jgm\" (UID: \"637b5bb3-b183-43dd-bca4-0349ac11e7d5\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.779711 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-dbus-socket\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.779768 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/796eb450-becc-4200-87e8-a9274d390901-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-h86r5\" (UID: \"796eb450-becc-4200-87e8-a9274d390901\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.779799 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkkxt\" (UniqueName: \"kubernetes.io/projected/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-kube-api-access-dkkxt\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.779830 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-ovs-socket\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.779872 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjq28\" (UniqueName: \"kubernetes.io/projected/796eb450-becc-4200-87e8-a9274d390901-kube-api-access-jjq28\") pod \"nmstate-webhook-5f6d4c5ccb-h86r5\" (UID: \"796eb450-becc-4200-87e8-a9274d390901\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.779889 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-nmstate-lock\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.799827 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9"] Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.800715 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.804073 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.804080 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.804295 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-2fmll" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.811223 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9"] Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881522 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-nmstate-lock\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjq28\" (UniqueName: \"kubernetes.io/projected/796eb450-becc-4200-87e8-a9274d390901-kube-api-access-jjq28\") pod \"nmstate-webhook-5f6d4c5ccb-h86r5\" (UID: \"796eb450-becc-4200-87e8-a9274d390901\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881633 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0648dfc-fcc2-49c9-863b-0c40c33ea417-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881686 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-nmstate-lock\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881789 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf924\" (UniqueName: \"kubernetes.io/projected/637b5bb3-b183-43dd-bca4-0349ac11e7d5-kube-api-access-vf924\") pod \"nmstate-metrics-7f946cbc9-52jgm\" (UID: \"637b5bb3-b183-43dd-bca4-0349ac11e7d5\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881814 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-dbus-socket\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/796eb450-becc-4200-87e8-a9274d390901-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-h86r5\" (UID: \"796eb450-becc-4200-87e8-a9274d390901\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881902 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4brv\" (UniqueName: \"kubernetes.io/projected/e0648dfc-fcc2-49c9-863b-0c40c33ea417-kube-api-access-b4brv\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881939 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkkxt\" (UniqueName: \"kubernetes.io/projected/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-kube-api-access-dkkxt\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.881970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-ovs-socket\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.882017 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e0648dfc-fcc2-49c9-863b-0c40c33ea417-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.882443 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-ovs-socket\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.882934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-dbus-socket\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.896414 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/796eb450-becc-4200-87e8-a9274d390901-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-h86r5\" (UID: \"796eb450-becc-4200-87e8-a9274d390901\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.900520 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjq28\" (UniqueName: \"kubernetes.io/projected/796eb450-becc-4200-87e8-a9274d390901-kube-api-access-jjq28\") pod \"nmstate-webhook-5f6d4c5ccb-h86r5\" (UID: \"796eb450-becc-4200-87e8-a9274d390901\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.902617 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf924\" (UniqueName: \"kubernetes.io/projected/637b5bb3-b183-43dd-bca4-0349ac11e7d5-kube-api-access-vf924\") pod \"nmstate-metrics-7f946cbc9-52jgm\" (UID: \"637b5bb3-b183-43dd-bca4-0349ac11e7d5\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.906186 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkkxt\" (UniqueName: \"kubernetes.io/projected/acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b-kube-api-access-dkkxt\") pod \"nmstate-handler-sccnl\" (UID: \"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b\") " pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.983625 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4brv\" (UniqueName: \"kubernetes.io/projected/e0648dfc-fcc2-49c9-863b-0c40c33ea417-kube-api-access-b4brv\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.984033 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e0648dfc-fcc2-49c9-863b-0c40c33ea417-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.984108 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0648dfc-fcc2-49c9-863b-0c40c33ea417-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:27 crc kubenswrapper[4809]: E1127 17:33:27.984229 4809 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 27 17:33:27 crc kubenswrapper[4809]: E1127 17:33:27.984279 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e0648dfc-fcc2-49c9-863b-0c40c33ea417-plugin-serving-cert podName:e0648dfc-fcc2-49c9-863b-0c40c33ea417 nodeName:}" failed. No retries permitted until 2025-11-27 17:33:28.48426319 +0000 UTC m=+1443.756720542 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/e0648dfc-fcc2-49c9-863b-0c40c33ea417-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-7lwv9" (UID: "e0648dfc-fcc2-49c9-863b-0c40c33ea417") : secret "plugin-serving-cert" not found Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.985060 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e0648dfc-fcc2-49c9-863b-0c40c33ea417-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:27 crc kubenswrapper[4809]: I1127 17:33:27.988194 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.000623 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6cf4cbb4d6-qrwhr"] Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.001913 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.004446 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.018639 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6cf4cbb4d6-qrwhr"] Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.033254 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.038682 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4brv\" (UniqueName: \"kubernetes.io/projected/e0648dfc-fcc2-49c9-863b-0c40c33ea417-kube-api-access-b4brv\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:28 crc kubenswrapper[4809]: W1127 17:33:28.059609 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacc7e45d_4ac0_43e7_a0ca_cd0ad3471f1b.slice/crio-bcd3cda0c2070159dc236d6947674142da014fe800739abf224e9a0f013eed8a WatchSource:0}: Error finding container bcd3cda0c2070159dc236d6947674142da014fe800739abf224e9a0f013eed8a: Status 404 returned error can't find the container with id bcd3cda0c2070159dc236d6947674142da014fe800739abf224e9a0f013eed8a Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.085252 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06557e94-7c1d-4f61-b817-ab6ab85578a3-console-serving-cert\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.085321 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nqwt\" (UniqueName: \"kubernetes.io/projected/06557e94-7c1d-4f61-b817-ab6ab85578a3-kube-api-access-6nqwt\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.085528 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-oauth-serving-cert\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.085578 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-service-ca\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.085601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06557e94-7c1d-4f61-b817-ab6ab85578a3-console-oauth-config\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.085659 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-console-config\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.085677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-trusted-ca-bundle\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.187337 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-oauth-serving-cert\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.187704 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-service-ca\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.187728 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06557e94-7c1d-4f61-b817-ab6ab85578a3-console-oauth-config\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.187771 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-console-config\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.187788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-trusted-ca-bundle\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.188894 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-service-ca\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.189489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-oauth-serving-cert\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.189609 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-console-config\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.187825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06557e94-7c1d-4f61-b817-ab6ab85578a3-console-serving-cert\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.189789 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nqwt\" (UniqueName: \"kubernetes.io/projected/06557e94-7c1d-4f61-b817-ab6ab85578a3-kube-api-access-6nqwt\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.189840 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06557e94-7c1d-4f61-b817-ab6ab85578a3-trusted-ca-bundle\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.198422 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06557e94-7c1d-4f61-b817-ab6ab85578a3-console-oauth-config\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.199229 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06557e94-7c1d-4f61-b817-ab6ab85578a3-console-serving-cert\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.207157 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nqwt\" (UniqueName: \"kubernetes.io/projected/06557e94-7c1d-4f61-b817-ab6ab85578a3-kube-api-access-6nqwt\") pod \"console-6cf4cbb4d6-qrwhr\" (UID: \"06557e94-7c1d-4f61-b817-ab6ab85578a3\") " pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.260192 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5"] Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.304186 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm"] Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.397133 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.501061 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0648dfc-fcc2-49c9-863b-0c40c33ea417-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.507154 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0648dfc-fcc2-49c9-863b-0c40c33ea417-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7lwv9\" (UID: \"e0648dfc-fcc2-49c9-863b-0c40c33ea417\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.715451 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.732157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-sccnl" event={"ID":"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b","Type":"ContainerStarted","Data":"bcd3cda0c2070159dc236d6947674142da014fe800739abf224e9a0f013eed8a"} Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.733097 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" event={"ID":"796eb450-becc-4200-87e8-a9274d390901","Type":"ContainerStarted","Data":"512640dc8071a6283f2c0f74a6bf964a3482e0ce75f118e6f8f358169d487951"} Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.736186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm" event={"ID":"637b5bb3-b183-43dd-bca4-0349ac11e7d5","Type":"ContainerStarted","Data":"8528c5dcf3efe7fb19aeee2adbe6b762735c552809bf090f5a48c5c672fb9ac8"} Nov 27 17:33:28 crc kubenswrapper[4809]: I1127 17:33:28.788401 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6cf4cbb4d6-qrwhr"] Nov 27 17:33:29 crc kubenswrapper[4809]: I1127 17:33:29.102140 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9"] Nov 27 17:33:29 crc kubenswrapper[4809]: I1127 17:33:29.743905 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" event={"ID":"e0648dfc-fcc2-49c9-863b-0c40c33ea417","Type":"ContainerStarted","Data":"9e094ee96eb7a427714fa9a98850b43127d4413267dcd7120bfa2f891acc3107"} Nov 27 17:33:29 crc kubenswrapper[4809]: I1127 17:33:29.745835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6cf4cbb4d6-qrwhr" event={"ID":"06557e94-7c1d-4f61-b817-ab6ab85578a3","Type":"ContainerStarted","Data":"78e75a6f53485f10bd3df2293e726686df379bc0c642be8484ee7b0875e0ad0e"} Nov 27 17:33:29 crc kubenswrapper[4809]: I1127 17:33:29.745935 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6cf4cbb4d6-qrwhr" event={"ID":"06557e94-7c1d-4f61-b817-ab6ab85578a3","Type":"ContainerStarted","Data":"b83b5a1e386295721d45612e9903183bd37351e2adbb2fc9e40ab5ff86fa5011"} Nov 27 17:33:29 crc kubenswrapper[4809]: I1127 17:33:29.769333 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6cf4cbb4d6-qrwhr" podStartSLOduration=2.769312696 podStartE2EDuration="2.769312696s" podCreationTimestamp="2025-11-27 17:33:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:33:29.764423442 +0000 UTC m=+1445.036880814" watchObservedRunningTime="2025-11-27 17:33:29.769312696 +0000 UTC m=+1445.041770048" Nov 27 17:33:31 crc kubenswrapper[4809]: I1127 17:33:31.762683 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-sccnl" event={"ID":"acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b","Type":"ContainerStarted","Data":"ef6719932e0794396e882e8e813f58a9bb067907664c8d1c74ede8e7a1baa078"} Nov 27 17:33:31 crc kubenswrapper[4809]: I1127 17:33:31.763582 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:31 crc kubenswrapper[4809]: I1127 17:33:31.765067 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" event={"ID":"796eb450-becc-4200-87e8-a9274d390901","Type":"ContainerStarted","Data":"1f029030afc0bebd4a7849e2ee8de61b956d55a5c04ff328be74fc754cb4dd61"} Nov 27 17:33:31 crc kubenswrapper[4809]: I1127 17:33:31.766509 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm" event={"ID":"637b5bb3-b183-43dd-bca4-0349ac11e7d5","Type":"ContainerStarted","Data":"a37d72e5e6fb08cd9e76fdf74b040c53e8ca3522c4a99c9d2f2e1da623051231"} Nov 27 17:33:31 crc kubenswrapper[4809]: I1127 17:33:31.766877 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:31 crc kubenswrapper[4809]: I1127 17:33:31.785502 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-sccnl" podStartSLOduration=2.256275101 podStartE2EDuration="4.785478421s" podCreationTimestamp="2025-11-27 17:33:27 +0000 UTC" firstStartedPulling="2025-11-27 17:33:28.062984913 +0000 UTC m=+1443.335442265" lastFinishedPulling="2025-11-27 17:33:30.592188233 +0000 UTC m=+1445.864645585" observedRunningTime="2025-11-27 17:33:31.779890979 +0000 UTC m=+1447.052348321" watchObservedRunningTime="2025-11-27 17:33:31.785478421 +0000 UTC m=+1447.057935773" Nov 27 17:33:31 crc kubenswrapper[4809]: I1127 17:33:31.799228 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" podStartSLOduration=2.471258762 podStartE2EDuration="4.799203964s" podCreationTimestamp="2025-11-27 17:33:27 +0000 UTC" firstStartedPulling="2025-11-27 17:33:28.266197443 +0000 UTC m=+1443.538654795" lastFinishedPulling="2025-11-27 17:33:30.594142655 +0000 UTC m=+1445.866599997" observedRunningTime="2025-11-27 17:33:31.792704757 +0000 UTC m=+1447.065162129" watchObservedRunningTime="2025-11-27 17:33:31.799203964 +0000 UTC m=+1447.071661316" Nov 27 17:33:32 crc kubenswrapper[4809]: I1127 17:33:32.773026 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" event={"ID":"e0648dfc-fcc2-49c9-863b-0c40c33ea417","Type":"ContainerStarted","Data":"992cf78c6d79a661fbf0eb45f78a6109d5827c2c229cf692ffd5e6e6964f8221"} Nov 27 17:33:32 crc kubenswrapper[4809]: I1127 17:33:32.794373 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7lwv9" podStartSLOduration=3.260100893 podStartE2EDuration="5.79435274s" podCreationTimestamp="2025-11-27 17:33:27 +0000 UTC" firstStartedPulling="2025-11-27 17:33:29.10890201 +0000 UTC m=+1444.381359362" lastFinishedPulling="2025-11-27 17:33:31.643153857 +0000 UTC m=+1446.915611209" observedRunningTime="2025-11-27 17:33:32.78847296 +0000 UTC m=+1448.060930332" watchObservedRunningTime="2025-11-27 17:33:32.79435274 +0000 UTC m=+1448.066810082" Nov 27 17:33:33 crc kubenswrapper[4809]: I1127 17:33:33.781498 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm" event={"ID":"637b5bb3-b183-43dd-bca4-0349ac11e7d5","Type":"ContainerStarted","Data":"03b082fb3f20c5dd00d66b9cc2fcdfac69a5c0b29cf3085316965162589ced59"} Nov 27 17:33:33 crc kubenswrapper[4809]: I1127 17:33:33.822575 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-52jgm" podStartSLOduration=1.8837140319999999 podStartE2EDuration="6.822550816s" podCreationTimestamp="2025-11-27 17:33:27 +0000 UTC" firstStartedPulling="2025-11-27 17:33:28.308346031 +0000 UTC m=+1443.580803383" lastFinishedPulling="2025-11-27 17:33:33.247182815 +0000 UTC m=+1448.519640167" observedRunningTime="2025-11-27 17:33:33.797398751 +0000 UTC m=+1449.069856133" watchObservedRunningTime="2025-11-27 17:33:33.822550816 +0000 UTC m=+1449.095008168" Nov 27 17:33:38 crc kubenswrapper[4809]: I1127 17:33:38.068572 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-sccnl" Nov 27 17:33:38 crc kubenswrapper[4809]: I1127 17:33:38.397377 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:38 crc kubenswrapper[4809]: I1127 17:33:38.397463 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:38 crc kubenswrapper[4809]: I1127 17:33:38.405897 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:38 crc kubenswrapper[4809]: I1127 17:33:38.819891 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6cf4cbb4d6-qrwhr" Nov 27 17:33:38 crc kubenswrapper[4809]: I1127 17:33:38.893211 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-w7f6q"] Nov 27 17:33:48 crc kubenswrapper[4809]: I1127 17:33:48.020667 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-h86r5" Nov 27 17:33:55 crc kubenswrapper[4809]: I1127 17:33:55.779343 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:33:55 crc kubenswrapper[4809]: I1127 17:33:55.779983 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:33:55 crc kubenswrapper[4809]: I1127 17:33:55.780041 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:33:55 crc kubenswrapper[4809]: I1127 17:33:55.780811 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"de4c3ebc223ee9f9580cd86c386eb6d52d4734548883eb67a85381672ea09fa4"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:33:55 crc kubenswrapper[4809]: I1127 17:33:55.780876 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://de4c3ebc223ee9f9580cd86c386eb6d52d4734548883eb67a85381672ea09fa4" gracePeriod=600 Nov 27 17:33:55 crc kubenswrapper[4809]: I1127 17:33:55.929439 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="de4c3ebc223ee9f9580cd86c386eb6d52d4734548883eb67a85381672ea09fa4" exitCode=0 Nov 27 17:33:55 crc kubenswrapper[4809]: I1127 17:33:55.929477 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"de4c3ebc223ee9f9580cd86c386eb6d52d4734548883eb67a85381672ea09fa4"} Nov 27 17:33:55 crc kubenswrapper[4809]: I1127 17:33:55.929507 4809 scope.go:117] "RemoveContainer" containerID="0f129562b4adcd0d1caac594ce6928b2856438351f75afb41cbf7f3b99003516" Nov 27 17:33:56 crc kubenswrapper[4809]: I1127 17:33:56.944921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596"} Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.277236 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf"] Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.279223 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.283232 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.287711 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf"] Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.334986 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.335078 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.335208 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-862wq\" (UniqueName: \"kubernetes.io/projected/11f39a67-042f-418d-9d4a-059243383431-kube-api-access-862wq\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.436248 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.436322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.436391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-862wq\" (UniqueName: \"kubernetes.io/projected/11f39a67-042f-418d-9d4a-059243383431-kube-api-access-862wq\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.436841 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.436856 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.458078 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-862wq\" (UniqueName: \"kubernetes.io/projected/11f39a67-042f-418d-9d4a-059243383431-kube-api-access-862wq\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.594880 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:02 crc kubenswrapper[4809]: I1127 17:34:02.993767 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf"] Nov 27 17:34:03 crc kubenswrapper[4809]: I1127 17:34:03.940968 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-w7f6q" podUID="5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" containerName="console" containerID="cri-o://179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c" gracePeriod=15 Nov 27 17:34:03 crc kubenswrapper[4809]: I1127 17:34:03.996818 4809 generic.go:334] "Generic (PLEG): container finished" podID="11f39a67-042f-418d-9d4a-059243383431" containerID="9b27a6e884bfbc6e243a0e07eca402bf35460d0116208b44c4a3c58c7bd3e317" exitCode=0 Nov 27 17:34:03 crc kubenswrapper[4809]: I1127 17:34:03.996866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" event={"ID":"11f39a67-042f-418d-9d4a-059243383431","Type":"ContainerDied","Data":"9b27a6e884bfbc6e243a0e07eca402bf35460d0116208b44c4a3c58c7bd3e317"} Nov 27 17:34:03 crc kubenswrapper[4809]: I1127 17:34:03.996893 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" event={"ID":"11f39a67-042f-418d-9d4a-059243383431","Type":"ContainerStarted","Data":"75278096604a4a7b0dfc7c2b77ce29fae1f31572f2ee3dc5c9b3b7230147386f"} Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.311571 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-w7f6q_5d5b9538-4c64-4cbb-b993-12a21ea1d0c7/console/0.log" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.311997 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.363674 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-oauth-serving-cert\") pod \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.363788 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-serving-cert\") pod \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.363818 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-trusted-ca-bundle\") pod \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.363844 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-config\") pod \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.363876 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-oauth-config\") pod \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.363914 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-service-ca\") pod \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.363965 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvtqt\" (UniqueName: \"kubernetes.io/projected/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-kube-api-access-kvtqt\") pod \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\" (UID: \"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7\") " Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.364723 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-config" (OuterVolumeSpecName: "console-config") pod "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" (UID: "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.364736 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-service-ca" (OuterVolumeSpecName: "service-ca") pod "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" (UID: "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.364839 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" (UID: "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.365150 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" (UID: "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.370684 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" (UID: "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.371015 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" (UID: "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.371023 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-kube-api-access-kvtqt" (OuterVolumeSpecName: "kube-api-access-kvtqt") pod "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" (UID: "5d5b9538-4c64-4cbb-b993-12a21ea1d0c7"). InnerVolumeSpecName "kube-api-access-kvtqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.465775 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.465806 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.465816 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.465825 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.465833 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.465841 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvtqt\" (UniqueName: \"kubernetes.io/projected/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-kube-api-access-kvtqt\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:04 crc kubenswrapper[4809]: I1127 17:34:04.465852 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.004551 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-w7f6q_5d5b9538-4c64-4cbb-b993-12a21ea1d0c7/console/0.log" Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.005784 4809 generic.go:334] "Generic (PLEG): container finished" podID="5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" containerID="179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c" exitCode=2 Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.005858 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-w7f6q" Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.005841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-w7f6q" event={"ID":"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7","Type":"ContainerDied","Data":"179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c"} Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.005976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-w7f6q" event={"ID":"5d5b9538-4c64-4cbb-b993-12a21ea1d0c7","Type":"ContainerDied","Data":"9451d2e23949e6519135565175511ad020967f65488d0e970ff5536a2fb77bec"} Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.005996 4809 scope.go:117] "RemoveContainer" containerID="179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c" Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.025874 4809 scope.go:117] "RemoveContainer" containerID="179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c" Nov 27 17:34:05 crc kubenswrapper[4809]: E1127 17:34:05.026328 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c\": container with ID starting with 179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c not found: ID does not exist" containerID="179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c" Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.026373 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c"} err="failed to get container status \"179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c\": rpc error: code = NotFound desc = could not find container \"179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c\": container with ID starting with 179313cabb3b98077649fc2c0f330b0d28275861b3abd234100febf077e5df3c not found: ID does not exist" Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.034348 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-w7f6q"] Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.038920 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-w7f6q"] Nov 27 17:34:05 crc kubenswrapper[4809]: I1127 17:34:05.466072 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" path="/var/lib/kubelet/pods/5d5b9538-4c64-4cbb-b993-12a21ea1d0c7/volumes" Nov 27 17:34:06 crc kubenswrapper[4809]: I1127 17:34:06.013288 4809 generic.go:334] "Generic (PLEG): container finished" podID="11f39a67-042f-418d-9d4a-059243383431" containerID="7085b16727edef97cd81d383616e8a56b1b0002af14bb0469017fcd2357a877a" exitCode=0 Nov 27 17:34:06 crc kubenswrapper[4809]: I1127 17:34:06.013331 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" event={"ID":"11f39a67-042f-418d-9d4a-059243383431","Type":"ContainerDied","Data":"7085b16727edef97cd81d383616e8a56b1b0002af14bb0469017fcd2357a877a"} Nov 27 17:34:07 crc kubenswrapper[4809]: I1127 17:34:07.021450 4809 generic.go:334] "Generic (PLEG): container finished" podID="11f39a67-042f-418d-9d4a-059243383431" containerID="86bba64fdcbbd73f53b8913424218cf0f7787cf656c64bf8b7fb50dad570b2e4" exitCode=0 Nov 27 17:34:07 crc kubenswrapper[4809]: I1127 17:34:07.021536 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" event={"ID":"11f39a67-042f-418d-9d4a-059243383431","Type":"ContainerDied","Data":"86bba64fdcbbd73f53b8913424218cf0f7787cf656c64bf8b7fb50dad570b2e4"} Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.283769 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.415776 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-bundle\") pod \"11f39a67-042f-418d-9d4a-059243383431\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.416093 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-util\") pod \"11f39a67-042f-418d-9d4a-059243383431\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.416147 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-862wq\" (UniqueName: \"kubernetes.io/projected/11f39a67-042f-418d-9d4a-059243383431-kube-api-access-862wq\") pod \"11f39a67-042f-418d-9d4a-059243383431\" (UID: \"11f39a67-042f-418d-9d4a-059243383431\") " Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.417170 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-bundle" (OuterVolumeSpecName: "bundle") pod "11f39a67-042f-418d-9d4a-059243383431" (UID: "11f39a67-042f-418d-9d4a-059243383431"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.429200 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11f39a67-042f-418d-9d4a-059243383431-kube-api-access-862wq" (OuterVolumeSpecName: "kube-api-access-862wq") pod "11f39a67-042f-418d-9d4a-059243383431" (UID: "11f39a67-042f-418d-9d4a-059243383431"). InnerVolumeSpecName "kube-api-access-862wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.432368 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-util" (OuterVolumeSpecName: "util") pod "11f39a67-042f-418d-9d4a-059243383431" (UID: "11f39a67-042f-418d-9d4a-059243383431"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.518115 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-util\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.518164 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-862wq\" (UniqueName: \"kubernetes.io/projected/11f39a67-042f-418d-9d4a-059243383431-kube-api-access-862wq\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:08 crc kubenswrapper[4809]: I1127 17:34:08.518186 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11f39a67-042f-418d-9d4a-059243383431-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:34:09 crc kubenswrapper[4809]: I1127 17:34:09.046270 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" event={"ID":"11f39a67-042f-418d-9d4a-059243383431","Type":"ContainerDied","Data":"75278096604a4a7b0dfc7c2b77ce29fae1f31572f2ee3dc5c9b3b7230147386f"} Nov 27 17:34:09 crc kubenswrapper[4809]: I1127 17:34:09.046312 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75278096604a4a7b0dfc7c2b77ce29fae1f31572f2ee3dc5c9b3b7230147386f" Nov 27 17:34:09 crc kubenswrapper[4809]: I1127 17:34:09.046374 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.293956 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7"] Nov 27 17:34:17 crc kubenswrapper[4809]: E1127 17:34:17.295055 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" containerName="console" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.295075 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" containerName="console" Nov 27 17:34:17 crc kubenswrapper[4809]: E1127 17:34:17.295091 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11f39a67-042f-418d-9d4a-059243383431" containerName="extract" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.295101 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="11f39a67-042f-418d-9d4a-059243383431" containerName="extract" Nov 27 17:34:17 crc kubenswrapper[4809]: E1127 17:34:17.295117 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11f39a67-042f-418d-9d4a-059243383431" containerName="util" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.295125 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="11f39a67-042f-418d-9d4a-059243383431" containerName="util" Nov 27 17:34:17 crc kubenswrapper[4809]: E1127 17:34:17.295144 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11f39a67-042f-418d-9d4a-059243383431" containerName="pull" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.295152 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="11f39a67-042f-418d-9d4a-059243383431" containerName="pull" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.295291 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d5b9538-4c64-4cbb-b993-12a21ea1d0c7" containerName="console" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.295312 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="11f39a67-042f-418d-9d4a-059243383431" containerName="extract" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.295963 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.297566 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8vb7f" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.298623 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.298920 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.299207 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.299910 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.319596 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7"] Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.439651 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thdl8\" (UniqueName: \"kubernetes.io/projected/eca3efdb-8eaf-4bd0-a2a4-471180514ee5-kube-api-access-thdl8\") pod \"metallb-operator-controller-manager-77cc889676-8xfh7\" (UID: \"eca3efdb-8eaf-4bd0-a2a4-471180514ee5\") " pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.439777 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eca3efdb-8eaf-4bd0-a2a4-471180514ee5-webhook-cert\") pod \"metallb-operator-controller-manager-77cc889676-8xfh7\" (UID: \"eca3efdb-8eaf-4bd0-a2a4-471180514ee5\") " pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.439826 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eca3efdb-8eaf-4bd0-a2a4-471180514ee5-apiservice-cert\") pod \"metallb-operator-controller-manager-77cc889676-8xfh7\" (UID: \"eca3efdb-8eaf-4bd0-a2a4-471180514ee5\") " pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.541878 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thdl8\" (UniqueName: \"kubernetes.io/projected/eca3efdb-8eaf-4bd0-a2a4-471180514ee5-kube-api-access-thdl8\") pod \"metallb-operator-controller-manager-77cc889676-8xfh7\" (UID: \"eca3efdb-8eaf-4bd0-a2a4-471180514ee5\") " pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.541914 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eca3efdb-8eaf-4bd0-a2a4-471180514ee5-webhook-cert\") pod \"metallb-operator-controller-manager-77cc889676-8xfh7\" (UID: \"eca3efdb-8eaf-4bd0-a2a4-471180514ee5\") " pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.541942 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eca3efdb-8eaf-4bd0-a2a4-471180514ee5-apiservice-cert\") pod \"metallb-operator-controller-manager-77cc889676-8xfh7\" (UID: \"eca3efdb-8eaf-4bd0-a2a4-471180514ee5\") " pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.549451 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eca3efdb-8eaf-4bd0-a2a4-471180514ee5-webhook-cert\") pod \"metallb-operator-controller-manager-77cc889676-8xfh7\" (UID: \"eca3efdb-8eaf-4bd0-a2a4-471180514ee5\") " pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.550409 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eca3efdb-8eaf-4bd0-a2a4-471180514ee5-apiservice-cert\") pod \"metallb-operator-controller-manager-77cc889676-8xfh7\" (UID: \"eca3efdb-8eaf-4bd0-a2a4-471180514ee5\") " pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.570662 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thdl8\" (UniqueName: \"kubernetes.io/projected/eca3efdb-8eaf-4bd0-a2a4-471180514ee5-kube-api-access-thdl8\") pod \"metallb-operator-controller-manager-77cc889676-8xfh7\" (UID: \"eca3efdb-8eaf-4bd0-a2a4-471180514ee5\") " pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.613590 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.650592 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf"] Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.651397 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.654013 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.654015 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.654097 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8rpl8" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.708580 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf"] Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.743789 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/779ed7ea-3841-4937-978f-046fe79f4a9c-apiservice-cert\") pod \"metallb-operator-webhook-server-5b59cf8cdf-wz5qf\" (UID: \"779ed7ea-3841-4937-978f-046fe79f4a9c\") " pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.743992 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj826\" (UniqueName: \"kubernetes.io/projected/779ed7ea-3841-4937-978f-046fe79f4a9c-kube-api-access-dj826\") pod \"metallb-operator-webhook-server-5b59cf8cdf-wz5qf\" (UID: \"779ed7ea-3841-4937-978f-046fe79f4a9c\") " pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.744215 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/779ed7ea-3841-4937-978f-046fe79f4a9c-webhook-cert\") pod \"metallb-operator-webhook-server-5b59cf8cdf-wz5qf\" (UID: \"779ed7ea-3841-4937-978f-046fe79f4a9c\") " pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.845182 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj826\" (UniqueName: \"kubernetes.io/projected/779ed7ea-3841-4937-978f-046fe79f4a9c-kube-api-access-dj826\") pod \"metallb-operator-webhook-server-5b59cf8cdf-wz5qf\" (UID: \"779ed7ea-3841-4937-978f-046fe79f4a9c\") " pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.845513 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/779ed7ea-3841-4937-978f-046fe79f4a9c-webhook-cert\") pod \"metallb-operator-webhook-server-5b59cf8cdf-wz5qf\" (UID: \"779ed7ea-3841-4937-978f-046fe79f4a9c\") " pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.845557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/779ed7ea-3841-4937-978f-046fe79f4a9c-apiservice-cert\") pod \"metallb-operator-webhook-server-5b59cf8cdf-wz5qf\" (UID: \"779ed7ea-3841-4937-978f-046fe79f4a9c\") " pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.849340 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/779ed7ea-3841-4937-978f-046fe79f4a9c-apiservice-cert\") pod \"metallb-operator-webhook-server-5b59cf8cdf-wz5qf\" (UID: \"779ed7ea-3841-4937-978f-046fe79f4a9c\") " pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.859155 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/779ed7ea-3841-4937-978f-046fe79f4a9c-webhook-cert\") pod \"metallb-operator-webhook-server-5b59cf8cdf-wz5qf\" (UID: \"779ed7ea-3841-4937-978f-046fe79f4a9c\") " pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.867368 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj826\" (UniqueName: \"kubernetes.io/projected/779ed7ea-3841-4937-978f-046fe79f4a9c-kube-api-access-dj826\") pod \"metallb-operator-webhook-server-5b59cf8cdf-wz5qf\" (UID: \"779ed7ea-3841-4937-978f-046fe79f4a9c\") " pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:17 crc kubenswrapper[4809]: I1127 17:34:17.990496 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:18 crc kubenswrapper[4809]: I1127 17:34:18.123952 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7"] Nov 27 17:34:18 crc kubenswrapper[4809]: W1127 17:34:18.137699 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeca3efdb_8eaf_4bd0_a2a4_471180514ee5.slice/crio-28497079d9e69f61c486403e510eb9f6f96eabf7506120e54462364ad853a210 WatchSource:0}: Error finding container 28497079d9e69f61c486403e510eb9f6f96eabf7506120e54462364ad853a210: Status 404 returned error can't find the container with id 28497079d9e69f61c486403e510eb9f6f96eabf7506120e54462364ad853a210 Nov 27 17:34:18 crc kubenswrapper[4809]: I1127 17:34:18.420654 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf"] Nov 27 17:34:18 crc kubenswrapper[4809]: W1127 17:34:18.424573 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod779ed7ea_3841_4937_978f_046fe79f4a9c.slice/crio-c34c44d4257cee91b81e9a582a3ab8b7040f9f14e7210748ae61d8f6cd8c3a90 WatchSource:0}: Error finding container c34c44d4257cee91b81e9a582a3ab8b7040f9f14e7210748ae61d8f6cd8c3a90: Status 404 returned error can't find the container with id c34c44d4257cee91b81e9a582a3ab8b7040f9f14e7210748ae61d8f6cd8c3a90 Nov 27 17:34:19 crc kubenswrapper[4809]: I1127 17:34:19.106925 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" event={"ID":"eca3efdb-8eaf-4bd0-a2a4-471180514ee5","Type":"ContainerStarted","Data":"28497079d9e69f61c486403e510eb9f6f96eabf7506120e54462364ad853a210"} Nov 27 17:34:19 crc kubenswrapper[4809]: I1127 17:34:19.108179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" event={"ID":"779ed7ea-3841-4937-978f-046fe79f4a9c","Type":"ContainerStarted","Data":"c34c44d4257cee91b81e9a582a3ab8b7040f9f14e7210748ae61d8f6cd8c3a90"} Nov 27 17:34:25 crc kubenswrapper[4809]: I1127 17:34:25.150730 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" event={"ID":"779ed7ea-3841-4937-978f-046fe79f4a9c","Type":"ContainerStarted","Data":"30d3270e3ff60d985482642c6ec6c4ae65ea58d1b9314ae2ce0b43dbf8452bce"} Nov 27 17:34:25 crc kubenswrapper[4809]: I1127 17:34:25.151337 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:25 crc kubenswrapper[4809]: I1127 17:34:25.152709 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" event={"ID":"eca3efdb-8eaf-4bd0-a2a4-471180514ee5","Type":"ContainerStarted","Data":"cbc63b1c83bee4e79f50d4445d58d779ad25aeacbee0d02c707f9b4440d21800"} Nov 27 17:34:25 crc kubenswrapper[4809]: I1127 17:34:25.152934 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:25 crc kubenswrapper[4809]: I1127 17:34:25.171695 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" podStartSLOduration=2.085171958 podStartE2EDuration="8.171667559s" podCreationTimestamp="2025-11-27 17:34:17 +0000 UTC" firstStartedPulling="2025-11-27 17:34:18.427517648 +0000 UTC m=+1493.699975000" lastFinishedPulling="2025-11-27 17:34:24.514013249 +0000 UTC m=+1499.786470601" observedRunningTime="2025-11-27 17:34:25.169239463 +0000 UTC m=+1500.441696815" watchObservedRunningTime="2025-11-27 17:34:25.171667559 +0000 UTC m=+1500.444124921" Nov 27 17:34:25 crc kubenswrapper[4809]: I1127 17:34:25.196319 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" podStartSLOduration=1.862628889 podStartE2EDuration="8.196298328s" podCreationTimestamp="2025-11-27 17:34:17 +0000 UTC" firstStartedPulling="2025-11-27 17:34:18.139691653 +0000 UTC m=+1493.412149005" lastFinishedPulling="2025-11-27 17:34:24.473361092 +0000 UTC m=+1499.745818444" observedRunningTime="2025-11-27 17:34:25.194471629 +0000 UTC m=+1500.466928981" watchObservedRunningTime="2025-11-27 17:34:25.196298328 +0000 UTC m=+1500.468755670" Nov 27 17:34:37 crc kubenswrapper[4809]: I1127 17:34:37.995258 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5b59cf8cdf-wz5qf" Nov 27 17:34:57 crc kubenswrapper[4809]: I1127 17:34:57.616367 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-77cc889676-8xfh7" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.318546 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-494tz"] Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.321429 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.323901 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.324770 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-5gsjw" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.324862 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.326509 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2"] Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.327531 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.329654 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.343255 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2"] Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.422140 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-gwn22"] Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.423798 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.430792 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-rbkql"] Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.431494 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.431540 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.431500 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.431869 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.432113 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-t8hmx" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.433940 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.444929 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-metrics\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.444988 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7p2q\" (UniqueName: \"kubernetes.io/projected/b7dfa097-6605-407b-8178-b4eb0497fd17-kube-api-access-p7p2q\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.445018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzsrn\" (UniqueName: \"kubernetes.io/projected/42ffbf32-2946-4087-8474-c0c0ecead8eb-kube-api-access-bzsrn\") pod \"frr-k8s-webhook-server-7fcb986d4-p48r2\" (UID: \"42ffbf32-2946-4087-8474-c0c0ecead8eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.445057 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b7dfa097-6605-407b-8178-b4eb0497fd17-frr-startup\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.445256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42ffbf32-2946-4087-8474-c0c0ecead8eb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-p48r2\" (UID: \"42ffbf32-2946-4087-8474-c0c0ecead8eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.445317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-frr-sockets\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.445364 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-reloader\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.445539 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-frr-conf\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.445590 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7dfa097-6605-407b-8178-b4eb0497fd17-metrics-certs\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.450787 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-rbkql"] Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547362 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42ffbf32-2946-4087-8474-c0c0ecead8eb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-p48r2\" (UID: \"42ffbf32-2946-4087-8474-c0c0ecead8eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-frr-sockets\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-reloader\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-frr-conf\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547544 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfw24\" (UniqueName: \"kubernetes.io/projected/976b64ef-1aff-4780-8353-dde2792907d4-kube-api-access-vfw24\") pod \"controller-f8648f98b-rbkql\" (UID: \"976b64ef-1aff-4780-8353-dde2792907d4\") " pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547592 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7dfa097-6605-407b-8178-b4eb0497fd17-metrics-certs\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547625 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/283ae992-ff54-437d-aba9-3dc45f78d876-metallb-excludel2\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/976b64ef-1aff-4780-8353-dde2792907d4-cert\") pod \"controller-f8648f98b-rbkql\" (UID: \"976b64ef-1aff-4780-8353-dde2792907d4\") " pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547711 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/976b64ef-1aff-4780-8353-dde2792907d4-metrics-certs\") pod \"controller-f8648f98b-rbkql\" (UID: \"976b64ef-1aff-4780-8353-dde2792907d4\") " pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547762 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-memberlist\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547922 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-metrics\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547945 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnmbh\" (UniqueName: \"kubernetes.io/projected/283ae992-ff54-437d-aba9-3dc45f78d876-kube-api-access-fnmbh\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7p2q\" (UniqueName: \"kubernetes.io/projected/b7dfa097-6605-407b-8178-b4eb0497fd17-kube-api-access-p7p2q\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.547999 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzsrn\" (UniqueName: \"kubernetes.io/projected/42ffbf32-2946-4087-8474-c0c0ecead8eb-kube-api-access-bzsrn\") pod \"frr-k8s-webhook-server-7fcb986d4-p48r2\" (UID: \"42ffbf32-2946-4087-8474-c0c0ecead8eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.548031 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-frr-conf\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.548089 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b7dfa097-6605-407b-8178-b4eb0497fd17-frr-startup\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.548099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-reloader\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.548113 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-metrics-certs\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.548392 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-frr-sockets\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.548435 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b7dfa097-6605-407b-8178-b4eb0497fd17-metrics\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.548885 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b7dfa097-6605-407b-8178-b4eb0497fd17-frr-startup\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.553382 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7dfa097-6605-407b-8178-b4eb0497fd17-metrics-certs\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.553455 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42ffbf32-2946-4087-8474-c0c0ecead8eb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-p48r2\" (UID: \"42ffbf32-2946-4087-8474-c0c0ecead8eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.566845 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7p2q\" (UniqueName: \"kubernetes.io/projected/b7dfa097-6605-407b-8178-b4eb0497fd17-kube-api-access-p7p2q\") pod \"frr-k8s-494tz\" (UID: \"b7dfa097-6605-407b-8178-b4eb0497fd17\") " pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.569011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzsrn\" (UniqueName: \"kubernetes.io/projected/42ffbf32-2946-4087-8474-c0c0ecead8eb-kube-api-access-bzsrn\") pod \"frr-k8s-webhook-server-7fcb986d4-p48r2\" (UID: \"42ffbf32-2946-4087-8474-c0c0ecead8eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.646446 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-494tz" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.649774 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/283ae992-ff54-437d-aba9-3dc45f78d876-metallb-excludel2\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.649841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/976b64ef-1aff-4780-8353-dde2792907d4-cert\") pod \"controller-f8648f98b-rbkql\" (UID: \"976b64ef-1aff-4780-8353-dde2792907d4\") " pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.649868 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/976b64ef-1aff-4780-8353-dde2792907d4-metrics-certs\") pod \"controller-f8648f98b-rbkql\" (UID: \"976b64ef-1aff-4780-8353-dde2792907d4\") " pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.649895 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-memberlist\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.649934 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnmbh\" (UniqueName: \"kubernetes.io/projected/283ae992-ff54-437d-aba9-3dc45f78d876-kube-api-access-fnmbh\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.649972 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-metrics-certs\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.650037 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfw24\" (UniqueName: \"kubernetes.io/projected/976b64ef-1aff-4780-8353-dde2792907d4-kube-api-access-vfw24\") pod \"controller-f8648f98b-rbkql\" (UID: \"976b64ef-1aff-4780-8353-dde2792907d4\") " pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: E1127 17:34:58.650161 4809 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 27 17:34:58 crc kubenswrapper[4809]: E1127 17:34:58.650213 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-memberlist podName:283ae992-ff54-437d-aba9-3dc45f78d876 nodeName:}" failed. No retries permitted until 2025-11-27 17:34:59.15019745 +0000 UTC m=+1534.422654802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-memberlist") pod "speaker-gwn22" (UID: "283ae992-ff54-437d-aba9-3dc45f78d876") : secret "metallb-memberlist" not found Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.650635 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/283ae992-ff54-437d-aba9-3dc45f78d876-metallb-excludel2\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.653309 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/976b64ef-1aff-4780-8353-dde2792907d4-metrics-certs\") pod \"controller-f8648f98b-rbkql\" (UID: \"976b64ef-1aff-4780-8353-dde2792907d4\") " pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.653563 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.654954 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-metrics-certs\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.657857 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.667569 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/976b64ef-1aff-4780-8353-dde2792907d4-cert\") pod \"controller-f8648f98b-rbkql\" (UID: \"976b64ef-1aff-4780-8353-dde2792907d4\") " pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.677178 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfw24\" (UniqueName: \"kubernetes.io/projected/976b64ef-1aff-4780-8353-dde2792907d4-kube-api-access-vfw24\") pod \"controller-f8648f98b-rbkql\" (UID: \"976b64ef-1aff-4780-8353-dde2792907d4\") " pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.680404 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnmbh\" (UniqueName: \"kubernetes.io/projected/283ae992-ff54-437d-aba9-3dc45f78d876-kube-api-access-fnmbh\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:58 crc kubenswrapper[4809]: I1127 17:34:58.754116 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:34:59 crc kubenswrapper[4809]: I1127 17:34:59.077445 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2"] Nov 27 17:34:59 crc kubenswrapper[4809]: W1127 17:34:59.080911 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42ffbf32_2946_4087_8474_c0c0ecead8eb.slice/crio-5c14376825b722eedde3859d44c0b1bd299ecf7ed75072083368a5c66ba1c7d0 WatchSource:0}: Error finding container 5c14376825b722eedde3859d44c0b1bd299ecf7ed75072083368a5c66ba1c7d0: Status 404 returned error can't find the container with id 5c14376825b722eedde3859d44c0b1bd299ecf7ed75072083368a5c66ba1c7d0 Nov 27 17:34:59 crc kubenswrapper[4809]: I1127 17:34:59.161663 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-memberlist\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:34:59 crc kubenswrapper[4809]: E1127 17:34:59.161895 4809 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 27 17:34:59 crc kubenswrapper[4809]: E1127 17:34:59.161984 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-memberlist podName:283ae992-ff54-437d-aba9-3dc45f78d876 nodeName:}" failed. No retries permitted until 2025-11-27 17:35:00.1619621 +0000 UTC m=+1535.434419452 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-memberlist") pod "speaker-gwn22" (UID: "283ae992-ff54-437d-aba9-3dc45f78d876") : secret "metallb-memberlist" not found Nov 27 17:34:59 crc kubenswrapper[4809]: I1127 17:34:59.177703 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-rbkql"] Nov 27 17:34:59 crc kubenswrapper[4809]: W1127 17:34:59.182804 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod976b64ef_1aff_4780_8353_dde2792907d4.slice/crio-6e0bc991f5cce1b1deeca9868e5be8133eed7881d1831a2482d1dbb186ac388c WatchSource:0}: Error finding container 6e0bc991f5cce1b1deeca9868e5be8133eed7881d1831a2482d1dbb186ac388c: Status 404 returned error can't find the container with id 6e0bc991f5cce1b1deeca9868e5be8133eed7881d1831a2482d1dbb186ac388c Nov 27 17:34:59 crc kubenswrapper[4809]: I1127 17:34:59.347936 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rbkql" event={"ID":"976b64ef-1aff-4780-8353-dde2792907d4","Type":"ContainerStarted","Data":"a91ab9eebb45000eff946583b6a5b4e959dbb64e44b1e72d244412823a71f864"} Nov 27 17:34:59 crc kubenswrapper[4809]: I1127 17:34:59.347984 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rbkql" event={"ID":"976b64ef-1aff-4780-8353-dde2792907d4","Type":"ContainerStarted","Data":"6e0bc991f5cce1b1deeca9868e5be8133eed7881d1831a2482d1dbb186ac388c"} Nov 27 17:34:59 crc kubenswrapper[4809]: I1127 17:34:59.349072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerStarted","Data":"70bbf17b3779450384b31d692cbdcd6bd1407e71b6bdbf3042890bb22590d528"} Nov 27 17:34:59 crc kubenswrapper[4809]: I1127 17:34:59.350072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" event={"ID":"42ffbf32-2946-4087-8474-c0c0ecead8eb","Type":"ContainerStarted","Data":"5c14376825b722eedde3859d44c0b1bd299ecf7ed75072083368a5c66ba1c7d0"} Nov 27 17:35:00 crc kubenswrapper[4809]: I1127 17:35:00.176094 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-memberlist\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:35:00 crc kubenswrapper[4809]: I1127 17:35:00.182930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/283ae992-ff54-437d-aba9-3dc45f78d876-memberlist\") pod \"speaker-gwn22\" (UID: \"283ae992-ff54-437d-aba9-3dc45f78d876\") " pod="metallb-system/speaker-gwn22" Nov 27 17:35:00 crc kubenswrapper[4809]: I1127 17:35:00.242732 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-gwn22" Nov 27 17:35:00 crc kubenswrapper[4809]: W1127 17:35:00.266943 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod283ae992_ff54_437d_aba9_3dc45f78d876.slice/crio-406544438c015fd0d4823a3b8674d5d457ed95e534f3c9282279339beaf2ac6a WatchSource:0}: Error finding container 406544438c015fd0d4823a3b8674d5d457ed95e534f3c9282279339beaf2ac6a: Status 404 returned error can't find the container with id 406544438c015fd0d4823a3b8674d5d457ed95e534f3c9282279339beaf2ac6a Nov 27 17:35:00 crc kubenswrapper[4809]: I1127 17:35:00.355760 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-gwn22" event={"ID":"283ae992-ff54-437d-aba9-3dc45f78d876","Type":"ContainerStarted","Data":"406544438c015fd0d4823a3b8674d5d457ed95e534f3c9282279339beaf2ac6a"} Nov 27 17:35:00 crc kubenswrapper[4809]: I1127 17:35:00.357185 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rbkql" event={"ID":"976b64ef-1aff-4780-8353-dde2792907d4","Type":"ContainerStarted","Data":"d002e836dd5cda320f0ba525bcf13fa4746a54c85d59a291d156c67974d05e84"} Nov 27 17:35:00 crc kubenswrapper[4809]: I1127 17:35:00.357699 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:35:00 crc kubenswrapper[4809]: I1127 17:35:00.374757 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-rbkql" podStartSLOduration=2.374722108 podStartE2EDuration="2.374722108s" podCreationTimestamp="2025-11-27 17:34:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:35:00.374385559 +0000 UTC m=+1535.646842911" watchObservedRunningTime="2025-11-27 17:35:00.374722108 +0000 UTC m=+1535.647179460" Nov 27 17:35:01 crc kubenswrapper[4809]: I1127 17:35:01.377390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-gwn22" event={"ID":"283ae992-ff54-437d-aba9-3dc45f78d876","Type":"ContainerStarted","Data":"3dfe36ce98b4d8fedee78f9d99bf2f4587d5ed4f1fbf7934654a9aecff66efcf"} Nov 27 17:35:01 crc kubenswrapper[4809]: I1127 17:35:01.377886 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-gwn22" event={"ID":"283ae992-ff54-437d-aba9-3dc45f78d876","Type":"ContainerStarted","Data":"37f74f06144d17e202b2e8e7d8ddcde1e762c271770bbde3a632c5ec5defc0a4"} Nov 27 17:35:01 crc kubenswrapper[4809]: I1127 17:35:01.397722 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-gwn22" podStartSLOduration=3.397704371 podStartE2EDuration="3.397704371s" podCreationTimestamp="2025-11-27 17:34:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:35:01.395599234 +0000 UTC m=+1536.668056596" watchObservedRunningTime="2025-11-27 17:35:01.397704371 +0000 UTC m=+1536.670161713" Nov 27 17:35:02 crc kubenswrapper[4809]: I1127 17:35:02.396000 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-gwn22" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.266882 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-psfk7"] Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.269101 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.274454 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-psfk7"] Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.359431 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cg4p\" (UniqueName: \"kubernetes.io/projected/4a667354-a328-4c06-b425-ae7e9da2826f-kube-api-access-2cg4p\") pod \"redhat-marketplace-psfk7\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.359508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-utilities\") pod \"redhat-marketplace-psfk7\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.359608 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-catalog-content\") pod \"redhat-marketplace-psfk7\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.460605 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-catalog-content\") pod \"redhat-marketplace-psfk7\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.460665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cg4p\" (UniqueName: \"kubernetes.io/projected/4a667354-a328-4c06-b425-ae7e9da2826f-kube-api-access-2cg4p\") pod \"redhat-marketplace-psfk7\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.460698 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-utilities\") pod \"redhat-marketplace-psfk7\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.461768 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-catalog-content\") pod \"redhat-marketplace-psfk7\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.462001 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-utilities\") pod \"redhat-marketplace-psfk7\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.502079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cg4p\" (UniqueName: \"kubernetes.io/projected/4a667354-a328-4c06-b425-ae7e9da2826f-kube-api-access-2cg4p\") pod \"redhat-marketplace-psfk7\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:04 crc kubenswrapper[4809]: I1127 17:35:04.596735 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:07 crc kubenswrapper[4809]: I1127 17:35:07.083561 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-psfk7"] Nov 27 17:35:07 crc kubenswrapper[4809]: I1127 17:35:07.447098 4809 generic.go:334] "Generic (PLEG): container finished" podID="b7dfa097-6605-407b-8178-b4eb0497fd17" containerID="7223a0c73246fde2f658bbc845a7df785ffa006162ec5d2111fa3b305f1ffe3c" exitCode=0 Nov 27 17:35:07 crc kubenswrapper[4809]: I1127 17:35:07.447151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerDied","Data":"7223a0c73246fde2f658bbc845a7df785ffa006162ec5d2111fa3b305f1ffe3c"} Nov 27 17:35:07 crc kubenswrapper[4809]: I1127 17:35:07.448807 4809 generic.go:334] "Generic (PLEG): container finished" podID="4a667354-a328-4c06-b425-ae7e9da2826f" containerID="62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f" exitCode=0 Nov 27 17:35:07 crc kubenswrapper[4809]: I1127 17:35:07.448861 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psfk7" event={"ID":"4a667354-a328-4c06-b425-ae7e9da2826f","Type":"ContainerDied","Data":"62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f"} Nov 27 17:35:07 crc kubenswrapper[4809]: I1127 17:35:07.448878 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psfk7" event={"ID":"4a667354-a328-4c06-b425-ae7e9da2826f","Type":"ContainerStarted","Data":"7ac70346dbabc8a2106835470c292741019b7f6edea7e6d007d894190134f956"} Nov 27 17:35:07 crc kubenswrapper[4809]: I1127 17:35:07.450704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" event={"ID":"42ffbf32-2946-4087-8474-c0c0ecead8eb","Type":"ContainerStarted","Data":"78a6d60e1632a1930664c9fa90deb989dc7a85c5653f4c9ecaaa539ab410d905"} Nov 27 17:35:07 crc kubenswrapper[4809]: I1127 17:35:07.451412 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:35:07 crc kubenswrapper[4809]: I1127 17:35:07.512858 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" podStartSLOduration=1.847461228 podStartE2EDuration="9.512841782s" podCreationTimestamp="2025-11-27 17:34:58 +0000 UTC" firstStartedPulling="2025-11-27 17:34:59.083125274 +0000 UTC m=+1534.355582626" lastFinishedPulling="2025-11-27 17:35:06.748505828 +0000 UTC m=+1542.020963180" observedRunningTime="2025-11-27 17:35:07.509647745 +0000 UTC m=+1542.782105097" watchObservedRunningTime="2025-11-27 17:35:07.512841782 +0000 UTC m=+1542.785299134" Nov 27 17:35:08 crc kubenswrapper[4809]: I1127 17:35:08.459779 4809 generic.go:334] "Generic (PLEG): container finished" podID="4a667354-a328-4c06-b425-ae7e9da2826f" containerID="e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273" exitCode=0 Nov 27 17:35:08 crc kubenswrapper[4809]: I1127 17:35:08.459873 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psfk7" event={"ID":"4a667354-a328-4c06-b425-ae7e9da2826f","Type":"ContainerDied","Data":"e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273"} Nov 27 17:35:08 crc kubenswrapper[4809]: I1127 17:35:08.461943 4809 generic.go:334] "Generic (PLEG): container finished" podID="b7dfa097-6605-407b-8178-b4eb0497fd17" containerID="d44f74e11b183c36d63a15f35ddbf4591bd91fa23182242c4fd115860ef6dbe0" exitCode=0 Nov 27 17:35:08 crc kubenswrapper[4809]: I1127 17:35:08.461997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerDied","Data":"d44f74e11b183c36d63a15f35ddbf4591bd91fa23182242c4fd115860ef6dbe0"} Nov 27 17:35:09 crc kubenswrapper[4809]: I1127 17:35:09.471564 4809 generic.go:334] "Generic (PLEG): container finished" podID="b7dfa097-6605-407b-8178-b4eb0497fd17" containerID="ea7d16a3e964b74a467a2b23a4221cf1588d1d28b03d8b3b9c1ac06aae65a4c0" exitCode=0 Nov 27 17:35:09 crc kubenswrapper[4809]: I1127 17:35:09.471614 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerDied","Data":"ea7d16a3e964b74a467a2b23a4221cf1588d1d28b03d8b3b9c1ac06aae65a4c0"} Nov 27 17:35:09 crc kubenswrapper[4809]: I1127 17:35:09.474930 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psfk7" event={"ID":"4a667354-a328-4c06-b425-ae7e9da2826f","Type":"ContainerStarted","Data":"d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8"} Nov 27 17:35:09 crc kubenswrapper[4809]: I1127 17:35:09.519965 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-psfk7" podStartSLOduration=3.949266249 podStartE2EDuration="5.51994639s" podCreationTimestamp="2025-11-27 17:35:04 +0000 UTC" firstStartedPulling="2025-11-27 17:35:07.449779745 +0000 UTC m=+1542.722237097" lastFinishedPulling="2025-11-27 17:35:09.020459886 +0000 UTC m=+1544.292917238" observedRunningTime="2025-11-27 17:35:09.515957452 +0000 UTC m=+1544.788414814" watchObservedRunningTime="2025-11-27 17:35:09.51994639 +0000 UTC m=+1544.792403742" Nov 27 17:35:10 crc kubenswrapper[4809]: I1127 17:35:10.246799 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-gwn22" Nov 27 17:35:10 crc kubenswrapper[4809]: I1127 17:35:10.486524 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerStarted","Data":"72e1fb6a4ce8c5b62ab76a1554e628ae9e3892788ed644e02fef58859e2430d6"} Nov 27 17:35:10 crc kubenswrapper[4809]: I1127 17:35:10.486565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerStarted","Data":"5649bbd8e1059e497f83c68c468abf0e5bf7da51d142d7df898ca24482b38d87"} Nov 27 17:35:10 crc kubenswrapper[4809]: I1127 17:35:10.486575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerStarted","Data":"7aa9abc8768ee2e4096f91e3288d2eb9c92f4c7da249e30ed592fa0d42508312"} Nov 27 17:35:10 crc kubenswrapper[4809]: I1127 17:35:10.486584 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerStarted","Data":"390760ab8f8e52a850c9dc490071e90c2aaf59f519a0a4f40a45a76d1963d319"} Nov 27 17:35:10 crc kubenswrapper[4809]: I1127 17:35:10.486592 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerStarted","Data":"da306ec9d554fdbaf86a3bc1ed593723e42df7ddc68b12cc3bc6665e0c6467ea"} Nov 27 17:35:11 crc kubenswrapper[4809]: I1127 17:35:11.495712 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-494tz" event={"ID":"b7dfa097-6605-407b-8178-b4eb0497fd17","Type":"ContainerStarted","Data":"7836601361d607a694c452f2f696d45ba8eff9b15c234e4758862acd876fd1b7"} Nov 27 17:35:11 crc kubenswrapper[4809]: I1127 17:35:11.496135 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-494tz" Nov 27 17:35:11 crc kubenswrapper[4809]: I1127 17:35:11.514427 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-494tz" podStartSLOduration=5.734434602 podStartE2EDuration="13.514408385s" podCreationTimestamp="2025-11-27 17:34:58 +0000 UTC" firstStartedPulling="2025-11-27 17:34:58.987118751 +0000 UTC m=+1534.259576093" lastFinishedPulling="2025-11-27 17:35:06.767092524 +0000 UTC m=+1542.039549876" observedRunningTime="2025-11-27 17:35:11.51311736 +0000 UTC m=+1546.785574712" watchObservedRunningTime="2025-11-27 17:35:11.514408385 +0000 UTC m=+1546.786865737" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.091000 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7d4jb"] Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.091972 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7d4jb" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.094344 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-t4fr5" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.094481 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.094435 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.121177 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8rjg\" (UniqueName: \"kubernetes.io/projected/60896b39-135c-407c-95ec-0b1c4aeed57d-kube-api-access-l8rjg\") pod \"openstack-operator-index-7d4jb\" (UID: \"60896b39-135c-407c-95ec-0b1c4aeed57d\") " pod="openstack-operators/openstack-operator-index-7d4jb" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.137508 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7d4jb"] Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.222460 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8rjg\" (UniqueName: \"kubernetes.io/projected/60896b39-135c-407c-95ec-0b1c4aeed57d-kube-api-access-l8rjg\") pod \"openstack-operator-index-7d4jb\" (UID: \"60896b39-135c-407c-95ec-0b1c4aeed57d\") " pod="openstack-operators/openstack-operator-index-7d4jb" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.244153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8rjg\" (UniqueName: \"kubernetes.io/projected/60896b39-135c-407c-95ec-0b1c4aeed57d-kube-api-access-l8rjg\") pod \"openstack-operator-index-7d4jb\" (UID: \"60896b39-135c-407c-95ec-0b1c4aeed57d\") " pod="openstack-operators/openstack-operator-index-7d4jb" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.425354 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7d4jb" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.647514 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-494tz" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.685945 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-494tz" Nov 27 17:35:13 crc kubenswrapper[4809]: I1127 17:35:13.830458 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7d4jb"] Nov 27 17:35:13 crc kubenswrapper[4809]: W1127 17:35:13.840066 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60896b39_135c_407c_95ec_0b1c4aeed57d.slice/crio-64d9f78680873e4688d3698a276c6a39131a620cc511af7bc438c9ce34dcdb9a WatchSource:0}: Error finding container 64d9f78680873e4688d3698a276c6a39131a620cc511af7bc438c9ce34dcdb9a: Status 404 returned error can't find the container with id 64d9f78680873e4688d3698a276c6a39131a620cc511af7bc438c9ce34dcdb9a Nov 27 17:35:14 crc kubenswrapper[4809]: I1127 17:35:14.517544 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7d4jb" event={"ID":"60896b39-135c-407c-95ec-0b1c4aeed57d","Type":"ContainerStarted","Data":"64d9f78680873e4688d3698a276c6a39131a620cc511af7bc438c9ce34dcdb9a"} Nov 27 17:35:14 crc kubenswrapper[4809]: I1127 17:35:14.597912 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:14 crc kubenswrapper[4809]: I1127 17:35:14.597958 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:14 crc kubenswrapper[4809]: I1127 17:35:14.643577 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:15 crc kubenswrapper[4809]: I1127 17:35:15.572156 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:16 crc kubenswrapper[4809]: I1127 17:35:16.657523 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7d4jb"] Nov 27 17:35:17 crc kubenswrapper[4809]: I1127 17:35:17.469952 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-ckqts"] Nov 27 17:35:17 crc kubenswrapper[4809]: I1127 17:35:17.471248 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ckqts" Nov 27 17:35:17 crc kubenswrapper[4809]: I1127 17:35:17.475229 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ckqts"] Nov 27 17:35:17 crc kubenswrapper[4809]: I1127 17:35:17.485914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hjl6\" (UniqueName: \"kubernetes.io/projected/45b7ef24-d97f-4dbe-8ad2-82f7f9983c68-kube-api-access-5hjl6\") pod \"openstack-operator-index-ckqts\" (UID: \"45b7ef24-d97f-4dbe-8ad2-82f7f9983c68\") " pod="openstack-operators/openstack-operator-index-ckqts" Nov 27 17:35:17 crc kubenswrapper[4809]: I1127 17:35:17.541935 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7d4jb" event={"ID":"60896b39-135c-407c-95ec-0b1c4aeed57d","Type":"ContainerStarted","Data":"648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f"} Nov 27 17:35:17 crc kubenswrapper[4809]: I1127 17:35:17.558652 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7d4jb" podStartSLOduration=2.006951635 podStartE2EDuration="4.558635046s" podCreationTimestamp="2025-11-27 17:35:13 +0000 UTC" firstStartedPulling="2025-11-27 17:35:13.842442199 +0000 UTC m=+1549.114899551" lastFinishedPulling="2025-11-27 17:35:16.39412561 +0000 UTC m=+1551.666582962" observedRunningTime="2025-11-27 17:35:17.556554859 +0000 UTC m=+1552.829012211" watchObservedRunningTime="2025-11-27 17:35:17.558635046 +0000 UTC m=+1552.831092398" Nov 27 17:35:17 crc kubenswrapper[4809]: I1127 17:35:17.587975 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hjl6\" (UniqueName: \"kubernetes.io/projected/45b7ef24-d97f-4dbe-8ad2-82f7f9983c68-kube-api-access-5hjl6\") pod \"openstack-operator-index-ckqts\" (UID: \"45b7ef24-d97f-4dbe-8ad2-82f7f9983c68\") " pod="openstack-operators/openstack-operator-index-ckqts" Nov 27 17:35:17 crc kubenswrapper[4809]: I1127 17:35:17.605607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hjl6\" (UniqueName: \"kubernetes.io/projected/45b7ef24-d97f-4dbe-8ad2-82f7f9983c68-kube-api-access-5hjl6\") pod \"openstack-operator-index-ckqts\" (UID: \"45b7ef24-d97f-4dbe-8ad2-82f7f9983c68\") " pod="openstack-operators/openstack-operator-index-ckqts" Nov 27 17:35:17 crc kubenswrapper[4809]: I1127 17:35:17.796412 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ckqts" Nov 27 17:35:18 crc kubenswrapper[4809]: I1127 17:35:18.196510 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ckqts"] Nov 27 17:35:18 crc kubenswrapper[4809]: W1127 17:35:18.202551 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45b7ef24_d97f_4dbe_8ad2_82f7f9983c68.slice/crio-dae3d3190253ea3ab887a62b2342826c36588b547ea49c9a5bb0baa58f65e350 WatchSource:0}: Error finding container dae3d3190253ea3ab887a62b2342826c36588b547ea49c9a5bb0baa58f65e350: Status 404 returned error can't find the container with id dae3d3190253ea3ab887a62b2342826c36588b547ea49c9a5bb0baa58f65e350 Nov 27 17:35:18 crc kubenswrapper[4809]: I1127 17:35:18.549797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ckqts" event={"ID":"45b7ef24-d97f-4dbe-8ad2-82f7f9983c68","Type":"ContainerStarted","Data":"055dee8c2bc485e0518f2ad25ae21d9f6b9a55f0194377ed44b19d389f59c49f"} Nov 27 17:35:18 crc kubenswrapper[4809]: I1127 17:35:18.550220 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ckqts" event={"ID":"45b7ef24-d97f-4dbe-8ad2-82f7f9983c68","Type":"ContainerStarted","Data":"dae3d3190253ea3ab887a62b2342826c36588b547ea49c9a5bb0baa58f65e350"} Nov 27 17:35:18 crc kubenswrapper[4809]: I1127 17:35:18.549882 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-7d4jb" podUID="60896b39-135c-407c-95ec-0b1c4aeed57d" containerName="registry-server" containerID="cri-o://648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f" gracePeriod=2 Nov 27 17:35:18 crc kubenswrapper[4809]: I1127 17:35:18.566772 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-ckqts" podStartSLOduration=1.5173222 podStartE2EDuration="1.566753385s" podCreationTimestamp="2025-11-27 17:35:17 +0000 UTC" firstStartedPulling="2025-11-27 17:35:18.206667684 +0000 UTC m=+1553.479125036" lastFinishedPulling="2025-11-27 17:35:18.256098869 +0000 UTC m=+1553.528556221" observedRunningTime="2025-11-27 17:35:18.562953952 +0000 UTC m=+1553.835411314" watchObservedRunningTime="2025-11-27 17:35:18.566753385 +0000 UTC m=+1553.839210737" Nov 27 17:35:18 crc kubenswrapper[4809]: I1127 17:35:18.674502 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-p48r2" Nov 27 17:35:18 crc kubenswrapper[4809]: I1127 17:35:18.758805 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-rbkql" Nov 27 17:35:18 crc kubenswrapper[4809]: I1127 17:35:18.981912 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7d4jb" Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.110964 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8rjg\" (UniqueName: \"kubernetes.io/projected/60896b39-135c-407c-95ec-0b1c4aeed57d-kube-api-access-l8rjg\") pod \"60896b39-135c-407c-95ec-0b1c4aeed57d\" (UID: \"60896b39-135c-407c-95ec-0b1c4aeed57d\") " Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.118233 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60896b39-135c-407c-95ec-0b1c4aeed57d-kube-api-access-l8rjg" (OuterVolumeSpecName: "kube-api-access-l8rjg") pod "60896b39-135c-407c-95ec-0b1c4aeed57d" (UID: "60896b39-135c-407c-95ec-0b1c4aeed57d"). InnerVolumeSpecName "kube-api-access-l8rjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.212513 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8rjg\" (UniqueName: \"kubernetes.io/projected/60896b39-135c-407c-95ec-0b1c4aeed57d-kube-api-access-l8rjg\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.456373 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-psfk7"] Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.456605 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-psfk7" podUID="4a667354-a328-4c06-b425-ae7e9da2826f" containerName="registry-server" containerID="cri-o://d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8" gracePeriod=2 Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.557095 4809 generic.go:334] "Generic (PLEG): container finished" podID="60896b39-135c-407c-95ec-0b1c4aeed57d" containerID="648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f" exitCode=0 Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.557806 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7d4jb" Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.558139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7d4jb" event={"ID":"60896b39-135c-407c-95ec-0b1c4aeed57d","Type":"ContainerDied","Data":"648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f"} Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.558166 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7d4jb" event={"ID":"60896b39-135c-407c-95ec-0b1c4aeed57d","Type":"ContainerDied","Data":"64d9f78680873e4688d3698a276c6a39131a620cc511af7bc438c9ce34dcdb9a"} Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.558182 4809 scope.go:117] "RemoveContainer" containerID="648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f" Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.579777 4809 scope.go:117] "RemoveContainer" containerID="648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f" Nov 27 17:35:19 crc kubenswrapper[4809]: E1127 17:35:19.580227 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f\": container with ID starting with 648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f not found: ID does not exist" containerID="648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f" Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.580262 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f"} err="failed to get container status \"648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f\": rpc error: code = NotFound desc = could not find container \"648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f\": container with ID starting with 648876146beede74ed33b46be89105d173aa2f24d1f40c6835421198d355f88f not found: ID does not exist" Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.585839 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7d4jb"] Nov 27 17:35:19 crc kubenswrapper[4809]: I1127 17:35:19.591305 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-7d4jb"] Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.273865 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.429335 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-utilities\") pod \"4a667354-a328-4c06-b425-ae7e9da2826f\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.430080 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-catalog-content\") pod \"4a667354-a328-4c06-b425-ae7e9da2826f\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.430123 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cg4p\" (UniqueName: \"kubernetes.io/projected/4a667354-a328-4c06-b425-ae7e9da2826f-kube-api-access-2cg4p\") pod \"4a667354-a328-4c06-b425-ae7e9da2826f\" (UID: \"4a667354-a328-4c06-b425-ae7e9da2826f\") " Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.430652 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-utilities" (OuterVolumeSpecName: "utilities") pod "4a667354-a328-4c06-b425-ae7e9da2826f" (UID: "4a667354-a328-4c06-b425-ae7e9da2826f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.437461 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a667354-a328-4c06-b425-ae7e9da2826f-kube-api-access-2cg4p" (OuterVolumeSpecName: "kube-api-access-2cg4p") pod "4a667354-a328-4c06-b425-ae7e9da2826f" (UID: "4a667354-a328-4c06-b425-ae7e9da2826f"). InnerVolumeSpecName "kube-api-access-2cg4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.452436 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a667354-a328-4c06-b425-ae7e9da2826f" (UID: "4a667354-a328-4c06-b425-ae7e9da2826f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.531369 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.531417 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a667354-a328-4c06-b425-ae7e9da2826f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.531433 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cg4p\" (UniqueName: \"kubernetes.io/projected/4a667354-a328-4c06-b425-ae7e9da2826f-kube-api-access-2cg4p\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.566757 4809 generic.go:334] "Generic (PLEG): container finished" podID="4a667354-a328-4c06-b425-ae7e9da2826f" containerID="d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8" exitCode=0 Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.566808 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-psfk7" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.566802 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psfk7" event={"ID":"4a667354-a328-4c06-b425-ae7e9da2826f","Type":"ContainerDied","Data":"d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8"} Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.566945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psfk7" event={"ID":"4a667354-a328-4c06-b425-ae7e9da2826f","Type":"ContainerDied","Data":"7ac70346dbabc8a2106835470c292741019b7f6edea7e6d007d894190134f956"} Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.566978 4809 scope.go:117] "RemoveContainer" containerID="d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.585967 4809 scope.go:117] "RemoveContainer" containerID="e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.598831 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-psfk7"] Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.603549 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-psfk7"] Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.624514 4809 scope.go:117] "RemoveContainer" containerID="62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.643197 4809 scope.go:117] "RemoveContainer" containerID="d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8" Nov 27 17:35:20 crc kubenswrapper[4809]: E1127 17:35:20.643770 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8\": container with ID starting with d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8 not found: ID does not exist" containerID="d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.643821 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8"} err="failed to get container status \"d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8\": rpc error: code = NotFound desc = could not find container \"d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8\": container with ID starting with d056933312c34bb72656443b2b20a4723e24e420ca15cfaff350322776789fd8 not found: ID does not exist" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.643852 4809 scope.go:117] "RemoveContainer" containerID="e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273" Nov 27 17:35:20 crc kubenswrapper[4809]: E1127 17:35:20.644311 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273\": container with ID starting with e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273 not found: ID does not exist" containerID="e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.644352 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273"} err="failed to get container status \"e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273\": rpc error: code = NotFound desc = could not find container \"e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273\": container with ID starting with e0d7d9b90321c4cc1ed4be81013cb729873d40a1e612d6faf5a8e96cf39e7273 not found: ID does not exist" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.644380 4809 scope.go:117] "RemoveContainer" containerID="62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f" Nov 27 17:35:20 crc kubenswrapper[4809]: E1127 17:35:20.644629 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f\": container with ID starting with 62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f not found: ID does not exist" containerID="62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f" Nov 27 17:35:20 crc kubenswrapper[4809]: I1127 17:35:20.644659 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f"} err="failed to get container status \"62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f\": rpc error: code = NotFound desc = could not find container \"62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f\": container with ID starting with 62a22dca5e49c4b61135b915389d454204b5b420c028737256cddb8953a3e44f not found: ID does not exist" Nov 27 17:35:21 crc kubenswrapper[4809]: I1127 17:35:21.466137 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a667354-a328-4c06-b425-ae7e9da2826f" path="/var/lib/kubelet/pods/4a667354-a328-4c06-b425-ae7e9da2826f/volumes" Nov 27 17:35:21 crc kubenswrapper[4809]: I1127 17:35:21.466735 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60896b39-135c-407c-95ec-0b1c4aeed57d" path="/var/lib/kubelet/pods/60896b39-135c-407c-95ec-0b1c4aeed57d/volumes" Nov 27 17:35:27 crc kubenswrapper[4809]: I1127 17:35:27.797491 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-ckqts" Nov 27 17:35:27 crc kubenswrapper[4809]: I1127 17:35:27.798112 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-ckqts" Nov 27 17:35:27 crc kubenswrapper[4809]: I1127 17:35:27.825375 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-ckqts" Nov 27 17:35:28 crc kubenswrapper[4809]: I1127 17:35:28.639186 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-ckqts" Nov 27 17:35:28 crc kubenswrapper[4809]: I1127 17:35:28.650128 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-494tz" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.907581 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq"] Nov 27 17:35:29 crc kubenswrapper[4809]: E1127 17:35:29.908006 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a667354-a328-4c06-b425-ae7e9da2826f" containerName="extract-utilities" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.908026 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a667354-a328-4c06-b425-ae7e9da2826f" containerName="extract-utilities" Nov 27 17:35:29 crc kubenswrapper[4809]: E1127 17:35:29.908048 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a667354-a328-4c06-b425-ae7e9da2826f" containerName="extract-content" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.908059 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a667354-a328-4c06-b425-ae7e9da2826f" containerName="extract-content" Nov 27 17:35:29 crc kubenswrapper[4809]: E1127 17:35:29.908080 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a667354-a328-4c06-b425-ae7e9da2826f" containerName="registry-server" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.908093 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a667354-a328-4c06-b425-ae7e9da2826f" containerName="registry-server" Nov 27 17:35:29 crc kubenswrapper[4809]: E1127 17:35:29.908108 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60896b39-135c-407c-95ec-0b1c4aeed57d" containerName="registry-server" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.908118 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="60896b39-135c-407c-95ec-0b1c4aeed57d" containerName="registry-server" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.908308 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a667354-a328-4c06-b425-ae7e9da2826f" containerName="registry-server" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.908328 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="60896b39-135c-407c-95ec-0b1c4aeed57d" containerName="registry-server" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.910760 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.912834 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-h2tdp" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.924635 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq"] Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.974211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzlht\" (UniqueName: \"kubernetes.io/projected/21484278-01d6-452c-806d-f8d364cb8325-kube-api-access-zzlht\") pod \"6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.974279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-util\") pod \"6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:29 crc kubenswrapper[4809]: I1127 17:35:29.974522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-bundle\") pod \"6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:30 crc kubenswrapper[4809]: I1127 17:35:30.078434 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-util\") pod \"6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:30 crc kubenswrapper[4809]: I1127 17:35:30.078898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-bundle\") pod \"6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:30 crc kubenswrapper[4809]: I1127 17:35:30.079342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzlht\" (UniqueName: \"kubernetes.io/projected/21484278-01d6-452c-806d-f8d364cb8325-kube-api-access-zzlht\") pod \"6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:30 crc kubenswrapper[4809]: I1127 17:35:30.080406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-util\") pod \"6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:30 crc kubenswrapper[4809]: I1127 17:35:30.080809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-bundle\") pod \"6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:30 crc kubenswrapper[4809]: I1127 17:35:30.102161 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzlht\" (UniqueName: \"kubernetes.io/projected/21484278-01d6-452c-806d-f8d364cb8325-kube-api-access-zzlht\") pod \"6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:30 crc kubenswrapper[4809]: I1127 17:35:30.233128 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:30 crc kubenswrapper[4809]: I1127 17:35:30.696697 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq"] Nov 27 17:35:31 crc kubenswrapper[4809]: I1127 17:35:31.635312 4809 generic.go:334] "Generic (PLEG): container finished" podID="21484278-01d6-452c-806d-f8d364cb8325" containerID="3392fcbc03806b02362377379fc3dee39b29b86d47ce67b75a6c060483661e10" exitCode=0 Nov 27 17:35:31 crc kubenswrapper[4809]: I1127 17:35:31.635367 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" event={"ID":"21484278-01d6-452c-806d-f8d364cb8325","Type":"ContainerDied","Data":"3392fcbc03806b02362377379fc3dee39b29b86d47ce67b75a6c060483661e10"} Nov 27 17:35:31 crc kubenswrapper[4809]: I1127 17:35:31.635645 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" event={"ID":"21484278-01d6-452c-806d-f8d364cb8325","Type":"ContainerStarted","Data":"cc038539ab9ca0afd715d2e731463973426a216d45e58cf67f1523a61d9c8147"} Nov 27 17:35:32 crc kubenswrapper[4809]: I1127 17:35:32.647100 4809 generic.go:334] "Generic (PLEG): container finished" podID="21484278-01d6-452c-806d-f8d364cb8325" containerID="34ce00b11d038fabb919d902172db6a4a848123cec365f33af9488998eded02c" exitCode=0 Nov 27 17:35:32 crc kubenswrapper[4809]: I1127 17:35:32.647344 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" event={"ID":"21484278-01d6-452c-806d-f8d364cb8325","Type":"ContainerDied","Data":"34ce00b11d038fabb919d902172db6a4a848123cec365f33af9488998eded02c"} Nov 27 17:35:33 crc kubenswrapper[4809]: I1127 17:35:33.657618 4809 generic.go:334] "Generic (PLEG): container finished" podID="21484278-01d6-452c-806d-f8d364cb8325" containerID="25506eb51046949f6bf741ddbaa7fa138004bd0bfa4f96b23c66443607d3beec" exitCode=0 Nov 27 17:35:33 crc kubenswrapper[4809]: I1127 17:35:33.657662 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" event={"ID":"21484278-01d6-452c-806d-f8d364cb8325","Type":"ContainerDied","Data":"25506eb51046949f6bf741ddbaa7fa138004bd0bfa4f96b23c66443607d3beec"} Nov 27 17:35:34 crc kubenswrapper[4809]: I1127 17:35:34.929447 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.066034 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzlht\" (UniqueName: \"kubernetes.io/projected/21484278-01d6-452c-806d-f8d364cb8325-kube-api-access-zzlht\") pod \"21484278-01d6-452c-806d-f8d364cb8325\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.066177 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-bundle\") pod \"21484278-01d6-452c-806d-f8d364cb8325\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.066247 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-util\") pod \"21484278-01d6-452c-806d-f8d364cb8325\" (UID: \"21484278-01d6-452c-806d-f8d364cb8325\") " Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.067886 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-bundle" (OuterVolumeSpecName: "bundle") pod "21484278-01d6-452c-806d-f8d364cb8325" (UID: "21484278-01d6-452c-806d-f8d364cb8325"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.071771 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21484278-01d6-452c-806d-f8d364cb8325-kube-api-access-zzlht" (OuterVolumeSpecName: "kube-api-access-zzlht") pod "21484278-01d6-452c-806d-f8d364cb8325" (UID: "21484278-01d6-452c-806d-f8d364cb8325"). InnerVolumeSpecName "kube-api-access-zzlht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.081488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-util" (OuterVolumeSpecName: "util") pod "21484278-01d6-452c-806d-f8d364cb8325" (UID: "21484278-01d6-452c-806d-f8d364cb8325"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.168717 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.168793 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21484278-01d6-452c-806d-f8d364cb8325-util\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.168804 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzlht\" (UniqueName: \"kubernetes.io/projected/21484278-01d6-452c-806d-f8d364cb8325-kube-api-access-zzlht\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.672859 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" event={"ID":"21484278-01d6-452c-806d-f8d364cb8325","Type":"ContainerDied","Data":"cc038539ab9ca0afd715d2e731463973426a216d45e58cf67f1523a61d9c8147"} Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.672910 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc038539ab9ca0afd715d2e731463973426a216d45e58cf67f1523a61d9c8147" Nov 27 17:35:35 crc kubenswrapper[4809]: I1127 17:35:35.672992 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.265383 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nmlrd"] Nov 27 17:35:36 crc kubenswrapper[4809]: E1127 17:35:36.265683 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21484278-01d6-452c-806d-f8d364cb8325" containerName="util" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.265699 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="21484278-01d6-452c-806d-f8d364cb8325" containerName="util" Nov 27 17:35:36 crc kubenswrapper[4809]: E1127 17:35:36.265718 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21484278-01d6-452c-806d-f8d364cb8325" containerName="pull" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.265724 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="21484278-01d6-452c-806d-f8d364cb8325" containerName="pull" Nov 27 17:35:36 crc kubenswrapper[4809]: E1127 17:35:36.265756 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21484278-01d6-452c-806d-f8d364cb8325" containerName="extract" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.265764 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="21484278-01d6-452c-806d-f8d364cb8325" containerName="extract" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.265904 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="21484278-01d6-452c-806d-f8d364cb8325" containerName="extract" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.266851 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.285694 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nmlrd"] Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.389569 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-utilities\") pod \"certified-operators-nmlrd\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.389619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-catalog-content\") pod \"certified-operators-nmlrd\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.389646 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mkfw\" (UniqueName: \"kubernetes.io/projected/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-kube-api-access-2mkfw\") pod \"certified-operators-nmlrd\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.490671 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-utilities\") pod \"certified-operators-nmlrd\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.490758 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-catalog-content\") pod \"certified-operators-nmlrd\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.490790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mkfw\" (UniqueName: \"kubernetes.io/projected/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-kube-api-access-2mkfw\") pod \"certified-operators-nmlrd\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.491241 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-utilities\") pod \"certified-operators-nmlrd\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.491360 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-catalog-content\") pod \"certified-operators-nmlrd\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.519012 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mkfw\" (UniqueName: \"kubernetes.io/projected/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-kube-api-access-2mkfw\") pod \"certified-operators-nmlrd\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:36 crc kubenswrapper[4809]: I1127 17:35:36.581827 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:37 crc kubenswrapper[4809]: I1127 17:35:37.012842 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nmlrd"] Nov 27 17:35:37 crc kubenswrapper[4809]: I1127 17:35:37.692654 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerID="4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040" exitCode=0 Nov 27 17:35:37 crc kubenswrapper[4809]: I1127 17:35:37.692813 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlrd" event={"ID":"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2","Type":"ContainerDied","Data":"4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040"} Nov 27 17:35:37 crc kubenswrapper[4809]: I1127 17:35:37.692962 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlrd" event={"ID":"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2","Type":"ContainerStarted","Data":"5f012056e92ace4ee69f48d22133d5a768a00d317aa82a2dfb4d6ecc92f6346d"} Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.447657 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd"] Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.449561 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.452219 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-qg6bs" Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.500964 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd"] Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.633196 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxwlk\" (UniqueName: \"kubernetes.io/projected/cceba3cb-7f9b-49d8-96e4-3e0e3927d106-kube-api-access-lxwlk\") pod \"openstack-operator-controller-operator-b44dff85c-jvvsd\" (UID: \"cceba3cb-7f9b-49d8-96e4-3e0e3927d106\") " pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.708389 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerID="33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144" exitCode=0 Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.708431 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlrd" event={"ID":"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2","Type":"ContainerDied","Data":"33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144"} Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.735047 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxwlk\" (UniqueName: \"kubernetes.io/projected/cceba3cb-7f9b-49d8-96e4-3e0e3927d106-kube-api-access-lxwlk\") pod \"openstack-operator-controller-operator-b44dff85c-jvvsd\" (UID: \"cceba3cb-7f9b-49d8-96e4-3e0e3927d106\") " pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.753606 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxwlk\" (UniqueName: \"kubernetes.io/projected/cceba3cb-7f9b-49d8-96e4-3e0e3927d106-kube-api-access-lxwlk\") pod \"openstack-operator-controller-operator-b44dff85c-jvvsd\" (UID: \"cceba3cb-7f9b-49d8-96e4-3e0e3927d106\") " pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" Nov 27 17:35:39 crc kubenswrapper[4809]: I1127 17:35:39.769124 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" Nov 27 17:35:40 crc kubenswrapper[4809]: I1127 17:35:40.191835 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd"] Nov 27 17:35:40 crc kubenswrapper[4809]: W1127 17:35:40.196942 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcceba3cb_7f9b_49d8_96e4_3e0e3927d106.slice/crio-19b1e217116840c704d601e2b4f4906a1317f933cbe3171ae2dd5d4a3fe7c270 WatchSource:0}: Error finding container 19b1e217116840c704d601e2b4f4906a1317f933cbe3171ae2dd5d4a3fe7c270: Status 404 returned error can't find the container with id 19b1e217116840c704d601e2b4f4906a1317f933cbe3171ae2dd5d4a3fe7c270 Nov 27 17:35:40 crc kubenswrapper[4809]: I1127 17:35:40.722318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" event={"ID":"cceba3cb-7f9b-49d8-96e4-3e0e3927d106","Type":"ContainerStarted","Data":"19b1e217116840c704d601e2b4f4906a1317f933cbe3171ae2dd5d4a3fe7c270"} Nov 27 17:35:40 crc kubenswrapper[4809]: I1127 17:35:40.731828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlrd" event={"ID":"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2","Type":"ContainerStarted","Data":"3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8"} Nov 27 17:35:40 crc kubenswrapper[4809]: I1127 17:35:40.759579 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nmlrd" podStartSLOduration=2.029461365 podStartE2EDuration="4.759557211s" podCreationTimestamp="2025-11-27 17:35:36 +0000 UTC" firstStartedPulling="2025-11-27 17:35:37.694100017 +0000 UTC m=+1572.966557359" lastFinishedPulling="2025-11-27 17:35:40.424195853 +0000 UTC m=+1575.696653205" observedRunningTime="2025-11-27 17:35:40.752150799 +0000 UTC m=+1576.024608171" watchObservedRunningTime="2025-11-27 17:35:40.759557211 +0000 UTC m=+1576.032014573" Nov 27 17:35:44 crc kubenswrapper[4809]: I1127 17:35:44.766342 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" event={"ID":"cceba3cb-7f9b-49d8-96e4-3e0e3927d106","Type":"ContainerStarted","Data":"6ee450dbb6231490850e5b3392721cb7777d7f48292104063f8ca234ac6573dd"} Nov 27 17:35:44 crc kubenswrapper[4809]: I1127 17:35:44.767019 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" Nov 27 17:35:44 crc kubenswrapper[4809]: I1127 17:35:44.805789 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" podStartSLOduration=2.001548607 podStartE2EDuration="5.805768589s" podCreationTimestamp="2025-11-27 17:35:39 +0000 UTC" firstStartedPulling="2025-11-27 17:35:40.199419516 +0000 UTC m=+1575.471876868" lastFinishedPulling="2025-11-27 17:35:44.003639498 +0000 UTC m=+1579.276096850" observedRunningTime="2025-11-27 17:35:44.804284049 +0000 UTC m=+1580.076741431" watchObservedRunningTime="2025-11-27 17:35:44.805768589 +0000 UTC m=+1580.078225961" Nov 27 17:35:46 crc kubenswrapper[4809]: I1127 17:35:46.582345 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:46 crc kubenswrapper[4809]: I1127 17:35:46.582705 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:46 crc kubenswrapper[4809]: I1127 17:35:46.622010 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:46 crc kubenswrapper[4809]: I1127 17:35:46.816783 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:47 crc kubenswrapper[4809]: I1127 17:35:47.464827 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nmlrd"] Nov 27 17:35:48 crc kubenswrapper[4809]: I1127 17:35:48.790005 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nmlrd" podUID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerName="registry-server" containerID="cri-o://3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8" gracePeriod=2 Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.226289 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.387585 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mkfw\" (UniqueName: \"kubernetes.io/projected/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-kube-api-access-2mkfw\") pod \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.387758 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-catalog-content\") pod \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.387875 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-utilities\") pod \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\" (UID: \"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2\") " Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.389360 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-utilities" (OuterVolumeSpecName: "utilities") pod "2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" (UID: "2c6cab54-04f0-4ab2-918e-5b45c94cb5f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.395010 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-kube-api-access-2mkfw" (OuterVolumeSpecName: "kube-api-access-2mkfw") pod "2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" (UID: "2c6cab54-04f0-4ab2-918e-5b45c94cb5f2"). InnerVolumeSpecName "kube-api-access-2mkfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.442113 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" (UID: "2c6cab54-04f0-4ab2-918e-5b45c94cb5f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.489441 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.489499 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mkfw\" (UniqueName: \"kubernetes.io/projected/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-kube-api-access-2mkfw\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.489514 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.771870 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-b44dff85c-jvvsd" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.799220 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerID="3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8" exitCode=0 Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.799277 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlrd" event={"ID":"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2","Type":"ContainerDied","Data":"3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8"} Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.799298 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmlrd" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.799323 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlrd" event={"ID":"2c6cab54-04f0-4ab2-918e-5b45c94cb5f2","Type":"ContainerDied","Data":"5f012056e92ace4ee69f48d22133d5a768a00d317aa82a2dfb4d6ecc92f6346d"} Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.799342 4809 scope.go:117] "RemoveContainer" containerID="3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.816328 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nmlrd"] Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.818034 4809 scope.go:117] "RemoveContainer" containerID="33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.821688 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nmlrd"] Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.834464 4809 scope.go:117] "RemoveContainer" containerID="4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.867147 4809 scope.go:117] "RemoveContainer" containerID="3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8" Nov 27 17:35:49 crc kubenswrapper[4809]: E1127 17:35:49.867674 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8\": container with ID starting with 3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8 not found: ID does not exist" containerID="3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.867713 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8"} err="failed to get container status \"3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8\": rpc error: code = NotFound desc = could not find container \"3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8\": container with ID starting with 3ef2166d308b318bba036d70ce5e4f59d4c63b1ab9352ea73c688f92d86556c8 not found: ID does not exist" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.867756 4809 scope.go:117] "RemoveContainer" containerID="33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144" Nov 27 17:35:49 crc kubenswrapper[4809]: E1127 17:35:49.868389 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144\": container with ID starting with 33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144 not found: ID does not exist" containerID="33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.868430 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144"} err="failed to get container status \"33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144\": rpc error: code = NotFound desc = could not find container \"33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144\": container with ID starting with 33ade371c4dfbf568ec48432fc22f3ab8f3c99729a4f29137f4aa01d8c0e5144 not found: ID does not exist" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.868463 4809 scope.go:117] "RemoveContainer" containerID="4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040" Nov 27 17:35:49 crc kubenswrapper[4809]: E1127 17:35:49.868826 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040\": container with ID starting with 4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040 not found: ID does not exist" containerID="4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040" Nov 27 17:35:49 crc kubenswrapper[4809]: I1127 17:35:49.868858 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040"} err="failed to get container status \"4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040\": rpc error: code = NotFound desc = could not find container \"4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040\": container with ID starting with 4592ce681fa114560ed17d6e6bcbe41129969810a9d36fe9bd1b51820aba7040 not found: ID does not exist" Nov 27 17:35:51 crc kubenswrapper[4809]: I1127 17:35:51.466474 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" path="/var/lib/kubelet/pods/2c6cab54-04f0-4ab2-918e-5b45c94cb5f2/volumes" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.597905 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5"] Nov 27 17:36:07 crc kubenswrapper[4809]: E1127 17:36:07.598873 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerName="registry-server" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.598891 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerName="registry-server" Nov 27 17:36:07 crc kubenswrapper[4809]: E1127 17:36:07.598911 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerName="extract-content" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.598919 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerName="extract-content" Nov 27 17:36:07 crc kubenswrapper[4809]: E1127 17:36:07.598936 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerName="extract-utilities" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.598944 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerName="extract-utilities" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.599095 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c6cab54-04f0-4ab2-918e-5b45c94cb5f2" containerName="registry-server" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.600030 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.602927 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.603990 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.604781 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-7k2z8" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.605980 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-4glrk" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.616793 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.627701 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-l58tv"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.629043 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.637894 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-7crf2" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.642777 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.652927 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-l58tv"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.673761 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.679959 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5789n\" (UniqueName: \"kubernetes.io/projected/be316447-6b83-4563-b75e-d4088882b695-kube-api-access-5789n\") pod \"cinder-operator-controller-manager-6b7f75547b-swx59\" (UID: \"be316447-6b83-4563-b75e-d4088882b695\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.680185 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5qvd\" (UniqueName: \"kubernetes.io/projected/4e5fbc76-cd41-432e-8c5b-90ff77d0704c-kube-api-access-f5qvd\") pod \"barbican-operator-controller-manager-7b64f4fb85-g7pm5\" (UID: \"4e5fbc76-cd41-432e-8c5b-90ff77d0704c\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.680236 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2lk4\" (UniqueName: \"kubernetes.io/projected/b470d686-15fd-4d31-ac6a-ffb60d3c9d79-kube-api-access-v2lk4\") pod \"designate-operator-controller-manager-955677c94-l58tv\" (UID: \"b470d686-15fd-4d31-ac6a-ffb60d3c9d79\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.682108 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.687125 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-8cwln" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.719636 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.735240 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.736477 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.743113 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-sl2l4" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.749378 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.750612 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.765070 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.770652 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-p75rw" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.782355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2lk4\" (UniqueName: \"kubernetes.io/projected/b470d686-15fd-4d31-ac6a-ffb60d3c9d79-kube-api-access-v2lk4\") pod \"designate-operator-controller-manager-955677c94-l58tv\" (UID: \"b470d686-15fd-4d31-ac6a-ffb60d3c9d79\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.782436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvvtf\" (UniqueName: \"kubernetes.io/projected/f6774dea-1e4a-4cd0-8688-5a588fe6d49b-kube-api-access-pvvtf\") pod \"glance-operator-controller-manager-589cbd6b5b-x4fdl\" (UID: \"f6774dea-1e4a-4cd0-8688-5a588fe6d49b\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.782461 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls7j7\" (UniqueName: \"kubernetes.io/projected/27f6291c-27aa-430f-864e-05be868c89b9-kube-api-access-ls7j7\") pod \"heat-operator-controller-manager-5b77f656f-4rlvc\" (UID: \"27f6291c-27aa-430f-864e-05be868c89b9\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.782522 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5789n\" (UniqueName: \"kubernetes.io/projected/be316447-6b83-4563-b75e-d4088882b695-kube-api-access-5789n\") pod \"cinder-operator-controller-manager-6b7f75547b-swx59\" (UID: \"be316447-6b83-4563-b75e-d4088882b695\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.782553 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5qvd\" (UniqueName: \"kubernetes.io/projected/4e5fbc76-cd41-432e-8c5b-90ff77d0704c-kube-api-access-f5qvd\") pod \"barbican-operator-controller-manager-7b64f4fb85-g7pm5\" (UID: \"4e5fbc76-cd41-432e-8c5b-90ff77d0704c\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.782575 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9mrh\" (UniqueName: \"kubernetes.io/projected/16fc4440-2960-4b7d-b059-c5192ea37a50-kube-api-access-n9mrh\") pod \"horizon-operator-controller-manager-5d494799bf-mkngh\" (UID: \"16fc4440-2960-4b7d-b059-c5192ea37a50\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.783873 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.810635 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-b28r8"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.812266 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.818152 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.818295 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vfclg" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.822447 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2lk4\" (UniqueName: \"kubernetes.io/projected/b470d686-15fd-4d31-ac6a-ffb60d3c9d79-kube-api-access-v2lk4\") pod \"designate-operator-controller-manager-955677c94-l58tv\" (UID: \"b470d686-15fd-4d31-ac6a-ffb60d3c9d79\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.822718 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5789n\" (UniqueName: \"kubernetes.io/projected/be316447-6b83-4563-b75e-d4088882b695-kube-api-access-5789n\") pod \"cinder-operator-controller-manager-6b7f75547b-swx59\" (UID: \"be316447-6b83-4563-b75e-d4088882b695\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.829108 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5qvd\" (UniqueName: \"kubernetes.io/projected/4e5fbc76-cd41-432e-8c5b-90ff77d0704c-kube-api-access-f5qvd\") pod \"barbican-operator-controller-manager-7b64f4fb85-g7pm5\" (UID: \"4e5fbc76-cd41-432e-8c5b-90ff77d0704c\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.832454 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-b28r8"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.854031 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.855227 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.858249 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-hfkpp" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.874993 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.876336 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.885250 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvvtf\" (UniqueName: \"kubernetes.io/projected/f6774dea-1e4a-4cd0-8688-5a588fe6d49b-kube-api-access-pvvtf\") pod \"glance-operator-controller-manager-589cbd6b5b-x4fdl\" (UID: \"f6774dea-1e4a-4cd0-8688-5a588fe6d49b\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.885292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls7j7\" (UniqueName: \"kubernetes.io/projected/27f6291c-27aa-430f-864e-05be868c89b9-kube-api-access-ls7j7\") pod \"heat-operator-controller-manager-5b77f656f-4rlvc\" (UID: \"27f6291c-27aa-430f-864e-05be868c89b9\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.885354 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9mrh\" (UniqueName: \"kubernetes.io/projected/16fc4440-2960-4b7d-b059-c5192ea37a50-kube-api-access-n9mrh\") pod \"horizon-operator-controller-manager-5d494799bf-mkngh\" (UID: \"16fc4440-2960-4b7d-b059-c5192ea37a50\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.885377 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkbmb\" (UniqueName: \"kubernetes.io/projected/976fdd57-7e54-4896-abca-914b86a98b19-kube-api-access-kkbmb\") pod \"ironic-operator-controller-manager-67cb4dc6d4-x9qq5\" (UID: \"976fdd57-7e54-4896-abca-914b86a98b19\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.885395 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx7bq\" (UniqueName: \"kubernetes.io/projected/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-kube-api-access-tx7bq\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.885422 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.885948 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-dmf9f" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.918218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9mrh\" (UniqueName: \"kubernetes.io/projected/16fc4440-2960-4b7d-b059-c5192ea37a50-kube-api-access-n9mrh\") pod \"horizon-operator-controller-manager-5d494799bf-mkngh\" (UID: \"16fc4440-2960-4b7d-b059-c5192ea37a50\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.919154 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.928226 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.928443 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls7j7\" (UniqueName: \"kubernetes.io/projected/27f6291c-27aa-430f-864e-05be868c89b9-kube-api-access-ls7j7\") pod \"heat-operator-controller-manager-5b77f656f-4rlvc\" (UID: \"27f6291c-27aa-430f-864e-05be868c89b9\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.932057 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvvtf\" (UniqueName: \"kubernetes.io/projected/f6774dea-1e4a-4cd0-8688-5a588fe6d49b-kube-api-access-pvvtf\") pod \"glance-operator-controller-manager-589cbd6b5b-x4fdl\" (UID: \"f6774dea-1e4a-4cd0-8688-5a588fe6d49b\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.960489 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.970825 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.988808 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8"] Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.990651 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" Nov 27 17:36:07 crc kubenswrapper[4809]: I1127 17:36:07.994975 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-46qnv" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:07.996874 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.002253 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n74b4\" (UniqueName: \"kubernetes.io/projected/63be6575-532b-487f-97a2-d9fd077e5be0-kube-api-access-n74b4\") pod \"keystone-operator-controller-manager-7b4567c7cf-mwv49\" (UID: \"63be6575-532b-487f-97a2-d9fd077e5be0\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.002553 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkbmb\" (UniqueName: \"kubernetes.io/projected/976fdd57-7e54-4896-abca-914b86a98b19-kube-api-access-kkbmb\") pod \"ironic-operator-controller-manager-67cb4dc6d4-x9qq5\" (UID: \"976fdd57-7e54-4896-abca-914b86a98b19\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.002668 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx7bq\" (UniqueName: \"kubernetes.io/projected/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-kube-api-access-tx7bq\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.000764 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.010097 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert podName:aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1 nodeName:}" failed. No retries permitted until 2025-11-27 17:36:08.51006981 +0000 UTC m=+1603.782527162 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert") pod "infra-operator-controller-manager-57548d458d-b28r8" (UID: "aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1") : secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.005364 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.010849 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.019037 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.020450 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.025315 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-fq7kj" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.025475 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.029384 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx7bq\" (UniqueName: \"kubernetes.io/projected/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-kube-api-access-tx7bq\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.034440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkbmb\" (UniqueName: \"kubernetes.io/projected/976fdd57-7e54-4896-abca-914b86a98b19-kube-api-access-kkbmb\") pod \"ironic-operator-controller-manager-67cb4dc6d4-x9qq5\" (UID: \"976fdd57-7e54-4896-abca-914b86a98b19\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.034928 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.057145 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.065926 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.070881 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.074863 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-wkmjc" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.080005 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.103163 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.103986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n74b4\" (UniqueName: \"kubernetes.io/projected/63be6575-532b-487f-97a2-d9fd077e5be0-kube-api-access-n74b4\") pod \"keystone-operator-controller-manager-7b4567c7cf-mwv49\" (UID: \"63be6575-532b-487f-97a2-d9fd077e5be0\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.104045 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tngfz\" (UniqueName: \"kubernetes.io/projected/9fb9364a-e568-418b-b403-dfa071e60297-kube-api-access-tngfz\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-tcwh6\" (UID: \"9fb9364a-e568-418b-b403-dfa071e60297\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.104081 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5zvb\" (UniqueName: \"kubernetes.io/projected/ad51c9a8-6543-403b-b8de-ff189da761a4-kube-api-access-d5zvb\") pod \"manila-operator-controller-manager-5d499bf58b-bzjd8\" (UID: \"ad51c9a8-6543-403b-b8de-ff189da761a4\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.104128 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5qg9\" (UniqueName: \"kubernetes.io/projected/db1b7072-b244-4e32-ba19-b8cd4254af2f-kube-api-access-s5qg9\") pod \"neutron-operator-controller-manager-6fdcddb789-6t8sk\" (UID: \"db1b7072-b244-4e32-ba19-b8cd4254af2f\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.116795 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.119251 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.122287 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-2vxhw" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.125376 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n74b4\" (UniqueName: \"kubernetes.io/projected/63be6575-532b-487f-97a2-d9fd077e5be0-kube-api-access-n74b4\") pod \"keystone-operator-controller-manager-7b4567c7cf-mwv49\" (UID: \"63be6575-532b-487f-97a2-d9fd077e5be0\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.127228 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.151155 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.154922 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.158148 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-784lj" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.205151 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln5pk\" (UniqueName: \"kubernetes.io/projected/1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659-kube-api-access-ln5pk\") pod \"octavia-operator-controller-manager-64cdc6ff96-xrjzd\" (UID: \"1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.205226 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5qg9\" (UniqueName: \"kubernetes.io/projected/db1b7072-b244-4e32-ba19-b8cd4254af2f-kube-api-access-s5qg9\") pod \"neutron-operator-controller-manager-6fdcddb789-6t8sk\" (UID: \"db1b7072-b244-4e32-ba19-b8cd4254af2f\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.205330 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p27z\" (UniqueName: \"kubernetes.io/projected/da239d06-b1c7-4b06-b830-5108030beed6-kube-api-access-5p27z\") pod \"nova-operator-controller-manager-79556f57fc-ll5l2\" (UID: \"da239d06-b1c7-4b06-b830-5108030beed6\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.205358 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tngfz\" (UniqueName: \"kubernetes.io/projected/9fb9364a-e568-418b-b403-dfa071e60297-kube-api-access-tngfz\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-tcwh6\" (UID: \"9fb9364a-e568-418b-b403-dfa071e60297\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.205405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5zvb\" (UniqueName: \"kubernetes.io/projected/ad51c9a8-6543-403b-b8de-ff189da761a4-kube-api-access-d5zvb\") pod \"manila-operator-controller-manager-5d499bf58b-bzjd8\" (UID: \"ad51c9a8-6543-403b-b8de-ff189da761a4\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.217844 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.218225 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.234160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5zvb\" (UniqueName: \"kubernetes.io/projected/ad51c9a8-6543-403b-b8de-ff189da761a4-kube-api-access-d5zvb\") pod \"manila-operator-controller-manager-5d499bf58b-bzjd8\" (UID: \"ad51c9a8-6543-403b-b8de-ff189da761a4\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.237504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tngfz\" (UniqueName: \"kubernetes.io/projected/9fb9364a-e568-418b-b403-dfa071e60297-kube-api-access-tngfz\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-tcwh6\" (UID: \"9fb9364a-e568-418b-b403-dfa071e60297\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.237654 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.242001 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.244898 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-kx586" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.245483 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.249910 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5qg9\" (UniqueName: \"kubernetes.io/projected/db1b7072-b244-4e32-ba19-b8cd4254af2f-kube-api-access-s5qg9\") pod \"neutron-operator-controller-manager-6fdcddb789-6t8sk\" (UID: \"db1b7072-b244-4e32-ba19-b8cd4254af2f\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.308844 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.352403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln5pk\" (UniqueName: \"kubernetes.io/projected/1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659-kube-api-access-ln5pk\") pod \"octavia-operator-controller-manager-64cdc6ff96-xrjzd\" (UID: \"1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.352525 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6brmx\" (UniqueName: \"kubernetes.io/projected/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-kube-api-access-6brmx\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.352552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.352580 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p27z\" (UniqueName: \"kubernetes.io/projected/da239d06-b1c7-4b06-b830-5108030beed6-kube-api-access-5p27z\") pod \"nova-operator-controller-manager-79556f57fc-ll5l2\" (UID: \"da239d06-b1c7-4b06-b830-5108030beed6\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.354684 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.372818 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.386341 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.401341 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.401485 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.410280 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.413366 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-b8jdx" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.425500 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p27z\" (UniqueName: \"kubernetes.io/projected/da239d06-b1c7-4b06-b830-5108030beed6-kube-api-access-5p27z\") pod \"nova-operator-controller-manager-79556f57fc-ll5l2\" (UID: \"da239d06-b1c7-4b06-b830-5108030beed6\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.431983 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln5pk\" (UniqueName: \"kubernetes.io/projected/1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659-kube-api-access-ln5pk\") pod \"octavia-operator-controller-manager-64cdc6ff96-xrjzd\" (UID: \"1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.445419 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.455230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6brmx\" (UniqueName: \"kubernetes.io/projected/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-kube-api-access-6brmx\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.455281 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.455338 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9nnj\" (UniqueName: \"kubernetes.io/projected/8ace4cdc-1253-4a0f-b667-83e997947f9b-kube-api-access-g9nnj\") pod \"ovn-operator-controller-manager-56897c768d-c9pds\" (UID: \"8ace4cdc-1253-4a0f-b667-83e997947f9b\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.455886 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.455950 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert podName:c077fc08-1ee2-493e-b631-f01ccd0b7c6e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:08.955927426 +0000 UTC m=+1604.228384778 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" (UID: "c077fc08-1ee2-493e-b631-f01ccd0b7c6e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.457563 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.495493 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.505436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6brmx\" (UniqueName: \"kubernetes.io/projected/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-kube-api-access-6brmx\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.522316 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.529823 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.536495 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-2hgcj" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.559908 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.562206 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.562204 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhzsk\" (UniqueName: \"kubernetes.io/projected/3da94235-386c-4f05-b869-2e08cee40d5d-kube-api-access-zhzsk\") pod \"placement-operator-controller-manager-57988cc5b5-8bq9l\" (UID: \"3da94235-386c-4f05-b869-2e08cee40d5d\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.565341 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9nnj\" (UniqueName: \"kubernetes.io/projected/8ace4cdc-1253-4a0f-b667-83e997947f9b-kube-api-access-g9nnj\") pod \"ovn-operator-controller-manager-56897c768d-c9pds\" (UID: \"8ace4cdc-1253-4a0f-b667-83e997947f9b\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.565388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.565574 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.565633 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert podName:aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1 nodeName:}" failed. No retries permitted until 2025-11-27 17:36:09.565605861 +0000 UTC m=+1604.838063213 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert") pod "infra-operator-controller-manager-57548d458d-b28r8" (UID: "aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1") : secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.568087 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-pc86j" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.572971 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.584546 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.586698 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9nnj\" (UniqueName: \"kubernetes.io/projected/8ace4cdc-1253-4a0f-b667-83e997947f9b-kube-api-access-g9nnj\") pod \"ovn-operator-controller-manager-56897c768d-c9pds\" (UID: \"8ace4cdc-1253-4a0f-b667-83e997947f9b\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.592026 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.593780 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.596764 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-5xpdr" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.633823 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.666387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hntlw\" (UniqueName: \"kubernetes.io/projected/6016f85d-a0f4-45ee-a96f-95dcb2c57e3c-kube-api-access-hntlw\") pod \"telemetry-operator-controller-manager-ff79b6df5-lq72j\" (UID: \"6016f85d-a0f4-45ee-a96f-95dcb2c57e3c\") " pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.666445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9gsf\" (UniqueName: \"kubernetes.io/projected/8f8e02f4-c34c-4b0a-b18f-8e088d48e12c-kube-api-access-p9gsf\") pod \"swift-operator-controller-manager-d77b94747-vvtmf\" (UID: \"8f8e02f4-c34c-4b0a-b18f-8e088d48e12c\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.666579 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhzsk\" (UniqueName: \"kubernetes.io/projected/3da94235-386c-4f05-b869-2e08cee40d5d-kube-api-access-zhzsk\") pod \"placement-operator-controller-manager-57988cc5b5-8bq9l\" (UID: \"3da94235-386c-4f05-b869-2e08cee40d5d\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.673532 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.674840 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.677080 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-wcrb9" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.688473 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhzsk\" (UniqueName: \"kubernetes.io/projected/3da94235-386c-4f05-b869-2e08cee40d5d-kube-api-access-zhzsk\") pod \"placement-operator-controller-manager-57988cc5b5-8bq9l\" (UID: \"3da94235-386c-4f05-b869-2e08cee40d5d\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.699958 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.728189 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.735337 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.741622 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-6m8gc" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.751714 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.768636 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv6rg\" (UniqueName: \"kubernetes.io/projected/8798a938-aa20-40e2-be26-408026fb123e-kube-api-access-jv6rg\") pod \"test-operator-controller-manager-5cd6c7f4c8-j6nv4\" (UID: \"8798a938-aa20-40e2-be26-408026fb123e\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.768703 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4d69\" (UniqueName: \"kubernetes.io/projected/5d49d329-6737-4bab-b243-64e618c0fa3b-kube-api-access-v4d69\") pod \"watcher-operator-controller-manager-656dcb59d4-d5hqw\" (UID: \"5d49d329-6737-4bab-b243-64e618c0fa3b\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.768777 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hntlw\" (UniqueName: \"kubernetes.io/projected/6016f85d-a0f4-45ee-a96f-95dcb2c57e3c-kube-api-access-hntlw\") pod \"telemetry-operator-controller-manager-ff79b6df5-lq72j\" (UID: \"6016f85d-a0f4-45ee-a96f-95dcb2c57e3c\") " pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.768806 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9gsf\" (UniqueName: \"kubernetes.io/projected/8f8e02f4-c34c-4b0a-b18f-8e088d48e12c-kube-api-access-p9gsf\") pod \"swift-operator-controller-manager-d77b94747-vvtmf\" (UID: \"8f8e02f4-c34c-4b0a-b18f-8e088d48e12c\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.774936 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.776731 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.779235 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.780884 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.781251 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-hkqks" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.788118 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.790037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hntlw\" (UniqueName: \"kubernetes.io/projected/6016f85d-a0f4-45ee-a96f-95dcb2c57e3c-kube-api-access-hntlw\") pod \"telemetry-operator-controller-manager-ff79b6df5-lq72j\" (UID: \"6016f85d-a0f4-45ee-a96f-95dcb2c57e3c\") " pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.794254 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.798311 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9gsf\" (UniqueName: \"kubernetes.io/projected/8f8e02f4-c34c-4b0a-b18f-8e088d48e12c-kube-api-access-p9gsf\") pod \"swift-operator-controller-manager-d77b94747-vvtmf\" (UID: \"8f8e02f4-c34c-4b0a-b18f-8e088d48e12c\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.800751 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.807442 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.809913 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-nfpjz" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.818393 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.830337 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp"] Nov 27 17:36:08 crc kubenswrapper[4809]: W1127 17:36:08.867848 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb470d686_15fd_4d31_ac6a_ffb60d3c9d79.slice/crio-b91a9e6ffeebd0f06b134e856513efef3e1274e0d2409cca34f3b908858f989f WatchSource:0}: Error finding container b91a9e6ffeebd0f06b134e856513efef3e1274e0d2409cca34f3b908858f989f: Status 404 returned error can't find the container with id b91a9e6ffeebd0f06b134e856513efef3e1274e0d2409cca34f3b908858f989f Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.869865 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.870015 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.870144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv6rg\" (UniqueName: \"kubernetes.io/projected/8798a938-aa20-40e2-be26-408026fb123e-kube-api-access-jv6rg\") pod \"test-operator-controller-manager-5cd6c7f4c8-j6nv4\" (UID: \"8798a938-aa20-40e2-be26-408026fb123e\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.870305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4d69\" (UniqueName: \"kubernetes.io/projected/5d49d329-6737-4bab-b243-64e618c0fa3b-kube-api-access-v4d69\") pod \"watcher-operator-controller-manager-656dcb59d4-d5hqw\" (UID: \"5d49d329-6737-4bab-b243-64e618c0fa3b\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.870406 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcv85\" (UniqueName: \"kubernetes.io/projected/21cb4ed4-198d-4dd0-b962-311bf2bcc850-kube-api-access-gcv85\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5b9vp\" (UID: \"21cb4ed4-198d-4dd0-b962-311bf2bcc850\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.870538 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-468bh\" (UniqueName: \"kubernetes.io/projected/18ae1240-d812-4f11-b515-6e7a94b8468e-kube-api-access-468bh\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.886511 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59"] Nov 27 17:36:08 crc kubenswrapper[4809]: W1127 17:36:08.887815 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6774dea_1e4a_4cd0_8688_5a588fe6d49b.slice/crio-31026cfef1e91aee62d8569a9262ccdbce8bfd638694a214ea2ed19958fa5b23 WatchSource:0}: Error finding container 31026cfef1e91aee62d8569a9262ccdbce8bfd638694a214ea2ed19958fa5b23: Status 404 returned error can't find the container with id 31026cfef1e91aee62d8569a9262ccdbce8bfd638694a214ea2ed19958fa5b23 Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.904462 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.912076 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4d69\" (UniqueName: \"kubernetes.io/projected/5d49d329-6737-4bab-b243-64e618c0fa3b-kube-api-access-v4d69\") pod \"watcher-operator-controller-manager-656dcb59d4-d5hqw\" (UID: \"5d49d329-6737-4bab-b243-64e618c0fa3b\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.927291 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.953106 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.959344 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv6rg\" (UniqueName: \"kubernetes.io/projected/8798a938-aa20-40e2-be26-408026fb123e-kube-api-access-jv6rg\") pod \"test-operator-controller-manager-5cd6c7f4c8-j6nv4\" (UID: \"8798a938-aa20-40e2-be26-408026fb123e\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.973021 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.973065 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.975912 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.975970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcv85\" (UniqueName: \"kubernetes.io/projected/21cb4ed4-198d-4dd0-b962-311bf2bcc850-kube-api-access-gcv85\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5b9vp\" (UID: \"21cb4ed4-198d-4dd0-b962-311bf2bcc850\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp" Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.976021 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-468bh\" (UniqueName: \"kubernetes.io/projected/18ae1240-d812-4f11-b515-6e7a94b8468e-kube-api-access-468bh\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.976449 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.976485 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert podName:c077fc08-1ee2-493e-b631-f01ccd0b7c6e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:09.976471853 +0000 UTC m=+1605.248929205 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" (UID: "c077fc08-1ee2-493e-b631-f01ccd0b7c6e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.976785 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.976810 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:09.476802293 +0000 UTC m=+1604.749259645 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "metrics-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.976841 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: E1127 17:36:08.976859 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:09.476853564 +0000 UTC m=+1604.749310916 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "webhook-server-cert" not found Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.978921 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-l58tv"] Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.992727 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" event={"ID":"4e5fbc76-cd41-432e-8c5b-90ff77d0704c","Type":"ContainerStarted","Data":"bcf1376f602df65748b62f4f51c7408c2183d9654ac8badec04912f2a2310ad8"} Nov 27 17:36:08 crc kubenswrapper[4809]: I1127 17:36:08.995766 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.000425 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" event={"ID":"b470d686-15fd-4d31-ac6a-ffb60d3c9d79","Type":"ContainerStarted","Data":"b91a9e6ffeebd0f06b134e856513efef3e1274e0d2409cca34f3b908858f989f"} Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.001699 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcv85\" (UniqueName: \"kubernetes.io/projected/21cb4ed4-198d-4dd0-b962-311bf2bcc850-kube-api-access-gcv85\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5b9vp\" (UID: \"21cb4ed4-198d-4dd0-b962-311bf2bcc850\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp" Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.002753 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" event={"ID":"be316447-6b83-4563-b75e-d4088882b695","Type":"ContainerStarted","Data":"83c36f31f455b30602dd0278a414754b30719431a31cab7b627765ca82567297"} Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.003404 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.006720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" event={"ID":"f6774dea-1e4a-4cd0-8688-5a588fe6d49b","Type":"ContainerStarted","Data":"31026cfef1e91aee62d8569a9262ccdbce8bfd638694a214ea2ed19958fa5b23"} Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.011682 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-468bh\" (UniqueName: \"kubernetes.io/projected/18ae1240-d812-4f11-b515-6e7a94b8468e-kube-api-access-468bh\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.046262 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.052799 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.100550 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.108272 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.108874 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16fc4440_2960_4b7d_b059_c5192ea37a50.slice/crio-e33bf4306b1eb1fc9b666a54384bbfcce6c13d7217e4c31e192e84189097a9f6 WatchSource:0}: Error finding container e33bf4306b1eb1fc9b666a54384bbfcce6c13d7217e4c31e192e84189097a9f6: Status 404 returned error can't find the container with id e33bf4306b1eb1fc9b666a54384bbfcce6c13d7217e4c31e192e84189097a9f6 Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.118461 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb1b7072_b244_4e32_ba19_b8cd4254af2f.slice/crio-5852c1ef805c2aee394fe21f8700cbaf72e31b1d98fc62484ee18a01b1bae8d5 WatchSource:0}: Error finding container 5852c1ef805c2aee394fe21f8700cbaf72e31b1d98fc62484ee18a01b1bae8d5: Status 404 returned error can't find the container with id 5852c1ef805c2aee394fe21f8700cbaf72e31b1d98fc62484ee18a01b1bae8d5 Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.257884 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod976fdd57_7e54_4896_abca_914b86a98b19.slice/crio-1149f6f08ff877093a0383dc249bff98daa4dc11147907ea389ed0bffca6d944 WatchSource:0}: Error finding container 1149f6f08ff877093a0383dc249bff98daa4dc11147907ea389ed0bffca6d944: Status 404 returned error can't find the container with id 1149f6f08ff877093a0383dc249bff98daa4dc11147907ea389ed0bffca6d944 Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.258697 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.292459 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp" Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.324812 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.340453 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49"] Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.481917 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ad10b2b_c3ec_4fa5_a828_1dfe7ce02659.slice/crio-b27e30b7b9c95c92d9d205438893662eea90dcea2e9e6890afb820f7afefb7cb WatchSource:0}: Error finding container b27e30b7b9c95c92d9d205438893662eea90dcea2e9e6890afb820f7afefb7cb: Status 404 returned error can't find the container with id b27e30b7b9c95c92d9d205438893662eea90dcea2e9e6890afb820f7afefb7cb Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.485342 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.488613 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.488688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.489941 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.490011 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:10.489995821 +0000 UTC m=+1605.762453173 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "metrics-server-cert" not found Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.490337 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.490482 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:10.490465693 +0000 UTC m=+1605.762923045 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "webhook-server-cert" not found Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.501040 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6"] Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.507352 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad51c9a8_6543_403b_b8de_ff189da761a4.slice/crio-1ad0c92a85964d5bc2651963d4614efca600b3497035e1ce36fb5b1b6401d15b WatchSource:0}: Error finding container 1ad0c92a85964d5bc2651963d4614efca600b3497035e1ce36fb5b1b6401d15b: Status 404 returned error can't find the container with id 1ad0c92a85964d5bc2651963d4614efca600b3497035e1ce36fb5b1b6401d15b Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.508136 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.590608 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.590829 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.591060 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert podName:aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1 nodeName:}" failed. No retries permitted until 2025-11-27 17:36:11.591030821 +0000 UTC m=+1606.863488173 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert") pod "infra-operator-controller-manager-57548d458d-b28r8" (UID: "aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1") : secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.631885 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds"] Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.635323 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f8e02f4_c34c_4b0a_b18f_8e088d48e12c.slice/crio-5b0c58baa60ef2f30315beed3a8843eebefafb336703989e7f6aa02343f09aaf WatchSource:0}: Error finding container 5b0c58baa60ef2f30315beed3a8843eebefafb336703989e7f6aa02343f09aaf: Status 404 returned error can't find the container with id 5b0c58baa60ef2f30315beed3a8843eebefafb336703989e7f6aa02343f09aaf Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.642007 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf"] Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.647116 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ace4cdc_1253_4a0f_b667_83e997947f9b.slice/crio-90bbd0d21705baa0c1110700058676a47d16b5be3d469d474b215229dc063301 WatchSource:0}: Error finding container 90bbd0d21705baa0c1110700058676a47d16b5be3d469d474b215229dc063301: Status 404 returned error can't find the container with id 90bbd0d21705baa0c1110700058676a47d16b5be3d469d474b215229dc063301 Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.714151 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.733833 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l"] Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.736601 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d49d329_6737_4bab_b243_64e618c0fa3b.slice/crio-5b56b426107ee49aee962096a56847e89375d8f8feafc9b97321b22dd9071308 WatchSource:0}: Error finding container 5b56b426107ee49aee962096a56847e89375d8f8feafc9b97321b22dd9071308: Status 404 returned error can't find the container with id 5b56b426107ee49aee962096a56847e89375d8f8feafc9b97321b22dd9071308 Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.738316 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.5:5001/openstack-k8s-operators/telemetry-operator:bf35154a77d3f7d42763b9d6bf295684481cdc52,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hntlw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-ff79b6df5-lq72j_openstack-operators(6016f85d-a0f4-45ee-a96f-95dcb2c57e3c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.740217 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hntlw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-ff79b6df5-lq72j_openstack-operators(6016f85d-a0f4-45ee-a96f-95dcb2c57e3c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.740572 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v4d69,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-d5hqw_openstack-operators(5d49d329-6737-4bab-b243-64e618c0fa3b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.741559 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhzsk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-8bq9l_openstack-operators(3da94235-386c-4f05-b869-2e08cee40d5d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.741635 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" podUID="6016f85d-a0f4-45ee-a96f-95dcb2c57e3c" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.744314 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v4d69,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-d5hqw_openstack-operators(5d49d329-6737-4bab-b243-64e618c0fa3b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.744403 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhzsk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-8bq9l_openstack-operators(3da94235-386c-4f05-b869-2e08cee40d5d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.745460 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" podUID="5d49d329-6737-4bab-b243-64e618c0fa3b" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.745511 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" podUID="3da94235-386c-4f05-b869-2e08cee40d5d" Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.747436 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw"] Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.752676 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4"] Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.765151 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8798a938_aa20_40e2_be26_408026fb123e.slice/crio-1e5f8d4ee1902208b6eae69e1d6e8e507dd3ee7571cd22ebfc67518c5915103a WatchSource:0}: Error finding container 1e5f8d4ee1902208b6eae69e1d6e8e507dd3ee7571cd22ebfc67518c5915103a: Status 404 returned error can't find the container with id 1e5f8d4ee1902208b6eae69e1d6e8e507dd3ee7571cd22ebfc67518c5915103a Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.768444 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jv6rg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-j6nv4_openstack-operators(8798a938-aa20-40e2-be26-408026fb123e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.771234 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jv6rg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-j6nv4_openstack-operators(8798a938-aa20-40e2-be26-408026fb123e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:09 crc kubenswrapper[4809]: E1127 17:36:09.773079 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" podUID="8798a938-aa20-40e2-be26-408026fb123e" Nov 27 17:36:09 crc kubenswrapper[4809]: I1127 17:36:09.866266 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp"] Nov 27 17:36:09 crc kubenswrapper[4809]: W1127 17:36:09.882342 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21cb4ed4_198d_4dd0_b962_311bf2bcc850.slice/crio-967617fe3bc9171745ec3559aee8bd85f873a74b5d2b8c47dd42834d28218562 WatchSource:0}: Error finding container 967617fe3bc9171745ec3559aee8bd85f873a74b5d2b8c47dd42834d28218562: Status 404 returned error can't find the container with id 967617fe3bc9171745ec3559aee8bd85f873a74b5d2b8c47dd42834d28218562 Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.001230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.001407 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.001488 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert podName:c077fc08-1ee2-493e-b631-f01ccd0b7c6e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:12.001469501 +0000 UTC m=+1607.273926853 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" (UID: "c077fc08-1ee2-493e-b631-f01ccd0b7c6e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.014514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" event={"ID":"16fc4440-2960-4b7d-b059-c5192ea37a50","Type":"ContainerStarted","Data":"e33bf4306b1eb1fc9b666a54384bbfcce6c13d7217e4c31e192e84189097a9f6"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.018141 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" event={"ID":"3da94235-386c-4f05-b869-2e08cee40d5d","Type":"ContainerStarted","Data":"fdf983729689ffb51229b2a4c052147ba815deac05bd800f082151393419fbd6"} Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.020296 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" podUID="3da94235-386c-4f05-b869-2e08cee40d5d" Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.022623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" event={"ID":"da239d06-b1c7-4b06-b830-5108030beed6","Type":"ContainerStarted","Data":"f1d3096ba54f8c44a77e4617d94e4894a0bafe260620a539715af5db38772254"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.030124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" event={"ID":"63be6575-532b-487f-97a2-d9fd077e5be0","Type":"ContainerStarted","Data":"64f9feeb16d859925da70ef55838622cacc3c0703c0be414986cb151c5b2b197"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.036049 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" event={"ID":"5d49d329-6737-4bab-b243-64e618c0fa3b","Type":"ContainerStarted","Data":"5b56b426107ee49aee962096a56847e89375d8f8feafc9b97321b22dd9071308"} Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.039657 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" podUID="5d49d329-6737-4bab-b243-64e618c0fa3b" Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.039955 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" event={"ID":"6016f85d-a0f4-45ee-a96f-95dcb2c57e3c","Type":"ContainerStarted","Data":"9f968b022a4a32579b9327d5c963baa5009996a7546a0cb916d5efbb3b788f59"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.047500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp" event={"ID":"21cb4ed4-198d-4dd0-b962-311bf2bcc850","Type":"ContainerStarted","Data":"967617fe3bc9171745ec3559aee8bd85f873a74b5d2b8c47dd42834d28218562"} Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.048916 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.5:5001/openstack-k8s-operators/telemetry-operator:bf35154a77d3f7d42763b9d6bf295684481cdc52\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" podUID="6016f85d-a0f4-45ee-a96f-95dcb2c57e3c" Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.048966 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" event={"ID":"8ace4cdc-1253-4a0f-b667-83e997947f9b","Type":"ContainerStarted","Data":"90bbd0d21705baa0c1110700058676a47d16b5be3d469d474b215229dc063301"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.050845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" event={"ID":"8f8e02f4-c34c-4b0a-b18f-8e088d48e12c","Type":"ContainerStarted","Data":"5b0c58baa60ef2f30315beed3a8843eebefafb336703989e7f6aa02343f09aaf"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.053633 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" event={"ID":"1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659","Type":"ContainerStarted","Data":"b27e30b7b9c95c92d9d205438893662eea90dcea2e9e6890afb820f7afefb7cb"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.057609 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" event={"ID":"976fdd57-7e54-4896-abca-914b86a98b19","Type":"ContainerStarted","Data":"1149f6f08ff877093a0383dc249bff98daa4dc11147907ea389ed0bffca6d944"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.062521 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" event={"ID":"db1b7072-b244-4e32-ba19-b8cd4254af2f","Type":"ContainerStarted","Data":"5852c1ef805c2aee394fe21f8700cbaf72e31b1d98fc62484ee18a01b1bae8d5"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.064730 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" event={"ID":"27f6291c-27aa-430f-864e-05be868c89b9","Type":"ContainerStarted","Data":"28bd461d573253a492ff086a4f29c16d9da4326fff94577097299ff9ff2bc79f"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.067413 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" event={"ID":"ad51c9a8-6543-403b-b8de-ff189da761a4","Type":"ContainerStarted","Data":"1ad0c92a85964d5bc2651963d4614efca600b3497035e1ce36fb5b1b6401d15b"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.072294 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" event={"ID":"9fb9364a-e568-418b-b403-dfa071e60297","Type":"ContainerStarted","Data":"5f65109bcb6daa03f961f2ce991d29e493a7187149e498000d976e8079ffc37f"} Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.075526 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" event={"ID":"8798a938-aa20-40e2-be26-408026fb123e","Type":"ContainerStarted","Data":"1e5f8d4ee1902208b6eae69e1d6e8e507dd3ee7571cd22ebfc67518c5915103a"} Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.080949 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" podUID="8798a938-aa20-40e2-be26-408026fb123e" Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.514036 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:10 crc kubenswrapper[4809]: I1127 17:36:10.514107 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.514262 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.514308 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:12.514293699 +0000 UTC m=+1607.786751051 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "webhook-server-cert" not found Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.514781 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 17:36:10 crc kubenswrapper[4809]: E1127 17:36:10.514823 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:12.514814794 +0000 UTC m=+1607.787272146 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "metrics-server-cert" not found Nov 27 17:36:11 crc kubenswrapper[4809]: E1127 17:36:11.094301 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" podUID="3da94235-386c-4f05-b869-2e08cee40d5d" Nov 27 17:36:11 crc kubenswrapper[4809]: E1127 17:36:11.094867 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.5:5001/openstack-k8s-operators/telemetry-operator:bf35154a77d3f7d42763b9d6bf295684481cdc52\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" podUID="6016f85d-a0f4-45ee-a96f-95dcb2c57e3c" Nov 27 17:36:11 crc kubenswrapper[4809]: E1127 17:36:11.095677 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" podUID="8798a938-aa20-40e2-be26-408026fb123e" Nov 27 17:36:11 crc kubenswrapper[4809]: E1127 17:36:11.097210 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" podUID="5d49d329-6737-4bab-b243-64e618c0fa3b" Nov 27 17:36:11 crc kubenswrapper[4809]: I1127 17:36:11.636092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:11 crc kubenswrapper[4809]: E1127 17:36:11.636109 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:11 crc kubenswrapper[4809]: E1127 17:36:11.636289 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert podName:aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1 nodeName:}" failed. No retries permitted until 2025-11-27 17:36:15.636268447 +0000 UTC m=+1610.908725799 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert") pod "infra-operator-controller-manager-57548d458d-b28r8" (UID: "aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1") : secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:12 crc kubenswrapper[4809]: I1127 17:36:12.044625 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:12 crc kubenswrapper[4809]: E1127 17:36:12.044824 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:12 crc kubenswrapper[4809]: E1127 17:36:12.044917 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert podName:c077fc08-1ee2-493e-b631-f01ccd0b7c6e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:16.044894009 +0000 UTC m=+1611.317351531 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" (UID: "c077fc08-1ee2-493e-b631-f01ccd0b7c6e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:12 crc kubenswrapper[4809]: I1127 17:36:12.552956 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:12 crc kubenswrapper[4809]: I1127 17:36:12.553074 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:12 crc kubenswrapper[4809]: E1127 17:36:12.553806 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 17:36:12 crc kubenswrapper[4809]: E1127 17:36:12.553851 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:16.553838051 +0000 UTC m=+1611.826295403 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "metrics-server-cert" not found Nov 27 17:36:12 crc kubenswrapper[4809]: E1127 17:36:12.554129 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 17:36:12 crc kubenswrapper[4809]: E1127 17:36:12.554163 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:16.55415361 +0000 UTC m=+1611.826610962 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "webhook-server-cert" not found Nov 27 17:36:15 crc kubenswrapper[4809]: I1127 17:36:15.707565 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:15 crc kubenswrapper[4809]: E1127 17:36:15.707829 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:15 crc kubenswrapper[4809]: E1127 17:36:15.708096 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert podName:aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1 nodeName:}" failed. No retries permitted until 2025-11-27 17:36:23.708074672 +0000 UTC m=+1618.980532024 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert") pod "infra-operator-controller-manager-57548d458d-b28r8" (UID: "aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1") : secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:16 crc kubenswrapper[4809]: I1127 17:36:16.112150 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:16 crc kubenswrapper[4809]: E1127 17:36:16.112725 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:16 crc kubenswrapper[4809]: E1127 17:36:16.112803 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert podName:c077fc08-1ee2-493e-b631-f01ccd0b7c6e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:24.112787528 +0000 UTC m=+1619.385244880 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" (UID: "c077fc08-1ee2-493e-b631-f01ccd0b7c6e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 17:36:16 crc kubenswrapper[4809]: I1127 17:36:16.621674 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:16 crc kubenswrapper[4809]: I1127 17:36:16.621748 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:16 crc kubenswrapper[4809]: E1127 17:36:16.621915 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 17:36:16 crc kubenswrapper[4809]: E1127 17:36:16.621952 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 17:36:16 crc kubenswrapper[4809]: E1127 17:36:16.622016 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:24.621990938 +0000 UTC m=+1619.894448300 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "metrics-server-cert" not found Nov 27 17:36:16 crc kubenswrapper[4809]: E1127 17:36:16.622046 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs podName:18ae1240-d812-4f11-b515-6e7a94b8468e nodeName:}" failed. No retries permitted until 2025-11-27 17:36:24.622036039 +0000 UTC m=+1619.894493581 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs") pod "openstack-operator-controller-manager-6644d5b8df-l9kv7" (UID: "18ae1240-d812-4f11-b515-6e7a94b8468e") : secret "webhook-server-cert" not found Nov 27 17:36:22 crc kubenswrapper[4809]: E1127 17:36:22.470149 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711" Nov 27 17:36:22 crc kubenswrapper[4809]: E1127 17:36:22.470938 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n74b4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7b4567c7cf-mwv49_openstack-operators(63be6575-532b-487f-97a2-d9fd077e5be0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.670869 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v2lk4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-955677c94-l58tv_openstack-operators(b470d686-15fd-4d31-ac6a-ffb60d3c9d79): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.673134 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" podUID="b470d686-15fd-4d31-ac6a-ffb60d3c9d79" Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.687927 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5p27z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-ll5l2_openstack-operators(da239d06-b1c7-4b06-b830-5108030beed6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.689528 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" podUID="da239d06-b1c7-4b06-b830-5108030beed6" Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.698707 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ln5pk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-64cdc6ff96-xrjzd_openstack-operators(1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.700002 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" podUID="1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659" Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.702062 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n9mrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5d494799bf-mkngh_openstack-operators(16fc4440-2960-4b7d-b059-c5192ea37a50): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.703289 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" podUID="16fc4440-2960-4b7d-b059-c5192ea37a50" Nov 27 17:36:23 crc kubenswrapper[4809]: I1127 17:36:23.749687 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.749942 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:23 crc kubenswrapper[4809]: E1127 17:36:23.750051 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert podName:aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1 nodeName:}" failed. No retries permitted until 2025-11-27 17:36:39.750018405 +0000 UTC m=+1635.022475757 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert") pod "infra-operator-controller-manager-57548d458d-b28r8" (UID: "aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1") : secret "infra-operator-webhook-server-cert" not found Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.155310 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.164785 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c077fc08-1ee2-493e-b631-f01ccd0b7c6e-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm\" (UID: \"c077fc08-1ee2-493e-b631-f01ccd0b7c6e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.186532 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" event={"ID":"b470d686-15fd-4d31-ac6a-ffb60d3c9d79","Type":"ContainerStarted","Data":"be3e1bf5421a152d12267cf0f5dc8551efcf3345cbeb0932eb9cb8fc77f6edd3"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.186821 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" Nov 27 17:36:24 crc kubenswrapper[4809]: E1127 17:36:24.190406 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" podUID="b470d686-15fd-4d31-ac6a-ffb60d3c9d79" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.204080 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" event={"ID":"976fdd57-7e54-4896-abca-914b86a98b19","Type":"ContainerStarted","Data":"76708796cd799586cf83a7af937adb0a1205ad218d531b57f4ce3475fe2b51c5"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.237308 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" event={"ID":"be316447-6b83-4563-b75e-d4088882b695","Type":"ContainerStarted","Data":"a3e1a0379879d22431df32513ecf7e0494981c7cf33b4686096b86f3570d29c4"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.263457 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp" event={"ID":"21cb4ed4-198d-4dd0-b962-311bf2bcc850","Type":"ContainerStarted","Data":"ccac5db820c3bf6125185b90edda9d3d3b6ed3c4395b57e4d3e16621b6bbaccb"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.272531 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.287422 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" event={"ID":"27f6291c-27aa-430f-864e-05be868c89b9","Type":"ContainerStarted","Data":"3d787fae11968e9b651e022e5a6c21269e7c6902ffb38373dc9b9366d2976912"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.329005 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" event={"ID":"f6774dea-1e4a-4cd0-8688-5a588fe6d49b","Type":"ContainerStarted","Data":"58bdd95b03bbeb1a8a91e40c0c38ce41cb0c2ffaf88f462f1f3d4493988f828e"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.387806 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5b9vp" podStartSLOduration=3.313723609 podStartE2EDuration="16.387774804s" podCreationTimestamp="2025-11-27 17:36:08 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.885396923 +0000 UTC m=+1605.157854275" lastFinishedPulling="2025-11-27 17:36:22.959448118 +0000 UTC m=+1618.231905470" observedRunningTime="2025-11-27 17:36:24.383210349 +0000 UTC m=+1619.655667701" watchObservedRunningTime="2025-11-27 17:36:24.387774804 +0000 UTC m=+1619.660232156" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.403029 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" event={"ID":"8ace4cdc-1253-4a0f-b667-83e997947f9b","Type":"ContainerStarted","Data":"f75418871af0c274575b6956a4802ebeec3b769f4b8767da0e51b1fe82e3eafe"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.444274 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" event={"ID":"8f8e02f4-c34c-4b0a-b18f-8e088d48e12c","Type":"ContainerStarted","Data":"daec2f88541de1806c45d753bda5458fc1a6dd2b959870ac288ca0c0a1eebaa9"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.459714 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" event={"ID":"ad51c9a8-6543-403b-b8de-ff189da761a4","Type":"ContainerStarted","Data":"69b9481409351637d6d26d63a0fc9a3045700a2f9d33581aca9d77af31ee92e1"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.474320 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" event={"ID":"9fb9364a-e568-418b-b403-dfa071e60297","Type":"ContainerStarted","Data":"84ee26e4f71a2c47daf2e4bd918446bb82298948c2c03b2dc3eefd7800fb6a67"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.485644 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" event={"ID":"16fc4440-2960-4b7d-b059-c5192ea37a50","Type":"ContainerStarted","Data":"6dbe0fbee1a6b5020071634dc166577ae4579db0005ab0cde703d97422f78d6a"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.485958 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" Nov 27 17:36:24 crc kubenswrapper[4809]: E1127 17:36:24.502972 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" podUID="16fc4440-2960-4b7d-b059-c5192ea37a50" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.519485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" event={"ID":"db1b7072-b244-4e32-ba19-b8cd4254af2f","Type":"ContainerStarted","Data":"548869114a7911c82f24afa2383ea0a0223ca8751e1febaed5a19df0aae1f9ef"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.539273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" event={"ID":"4e5fbc76-cd41-432e-8c5b-90ff77d0704c","Type":"ContainerStarted","Data":"c33bcfedd861ecc718213c92829e767aa14ae827b30071a501e45fda9f695e4c"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.564055 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" event={"ID":"da239d06-b1c7-4b06-b830-5108030beed6","Type":"ContainerStarted","Data":"e08534ed0039a33bcce770fd46cef073ede58db417749b0d4012de00753c1030"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.564660 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" Nov 27 17:36:24 crc kubenswrapper[4809]: E1127 17:36:24.565171 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" podUID="da239d06-b1c7-4b06-b830-5108030beed6" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.566370 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" event={"ID":"1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659","Type":"ContainerStarted","Data":"d6a5bdd225045b83001f3d88d5bfb380be0b5ab8e95c0281fb6ce82fc5fb41f3"} Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.566911 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" Nov 27 17:36:24 crc kubenswrapper[4809]: E1127 17:36:24.567802 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" podUID="1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.689791 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.689982 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.704687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-metrics-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.768850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18ae1240-d812-4f11-b515-6e7a94b8468e-webhook-certs\") pod \"openstack-operator-controller-manager-6644d5b8df-l9kv7\" (UID: \"18ae1240-d812-4f11-b515-6e7a94b8468e\") " pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.854114 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:24 crc kubenswrapper[4809]: I1127 17:36:24.929827 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm"] Nov 27 17:36:25 crc kubenswrapper[4809]: I1127 17:36:25.591351 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" event={"ID":"c077fc08-1ee2-493e-b631-f01ccd0b7c6e","Type":"ContainerStarted","Data":"6dae90f0cd4a3c0e943a27224832982ff84ee70332b3bec34a3382a8b34a05a5"} Nov 27 17:36:25 crc kubenswrapper[4809]: E1127 17:36:25.594295 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" podUID="da239d06-b1c7-4b06-b830-5108030beed6" Nov 27 17:36:25 crc kubenswrapper[4809]: E1127 17:36:25.594383 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" podUID="b470d686-15fd-4d31-ac6a-ffb60d3c9d79" Nov 27 17:36:25 crc kubenswrapper[4809]: E1127 17:36:25.594513 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" podUID="16fc4440-2960-4b7d-b059-c5192ea37a50" Nov 27 17:36:25 crc kubenswrapper[4809]: E1127 17:36:25.614967 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" podUID="1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659" Nov 27 17:36:25 crc kubenswrapper[4809]: I1127 17:36:25.784673 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:36:25 crc kubenswrapper[4809]: I1127 17:36:25.785419 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:36:26 crc kubenswrapper[4809]: I1127 17:36:26.008816 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7"] Nov 27 17:36:27 crc kubenswrapper[4809]: W1127 17:36:27.538807 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18ae1240_d812_4f11_b515_6e7a94b8468e.slice/crio-f8e31842e789911db0d2529b09698fb347c68a107f11160e63d6037c00087f6d WatchSource:0}: Error finding container f8e31842e789911db0d2529b09698fb347c68a107f11160e63d6037c00087f6d: Status 404 returned error can't find the container with id f8e31842e789911db0d2529b09698fb347c68a107f11160e63d6037c00087f6d Nov 27 17:36:27 crc kubenswrapper[4809]: I1127 17:36:27.608597 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" event={"ID":"18ae1240-d812-4f11-b515-6e7a94b8468e","Type":"ContainerStarted","Data":"f8e31842e789911db0d2529b09698fb347c68a107f11160e63d6037c00087f6d"} Nov 27 17:36:28 crc kubenswrapper[4809]: I1127 17:36:28.083848 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" Nov 27 17:36:28 crc kubenswrapper[4809]: E1127 17:36:28.088295 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" podUID="16fc4440-2960-4b7d-b059-c5192ea37a50" Nov 27 17:36:28 crc kubenswrapper[4809]: I1127 17:36:28.461005 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" Nov 27 17:36:28 crc kubenswrapper[4809]: I1127 17:36:28.500204 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" Nov 27 17:36:28 crc kubenswrapper[4809]: E1127 17:36:28.596858 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" podUID="1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659" Nov 27 17:36:28 crc kubenswrapper[4809]: E1127 17:36:28.597385 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" podUID="da239d06-b1c7-4b06-b830-5108030beed6" Nov 27 17:36:32 crc kubenswrapper[4809]: I1127 17:36:32.649415 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" event={"ID":"18ae1240-d812-4f11-b515-6e7a94b8468e","Type":"ContainerStarted","Data":"f2c299249ff13209a4eedfd5a8b9a6fa7d6ae8cf64e7ed2220899e7da0684489"} Nov 27 17:36:32 crc kubenswrapper[4809]: I1127 17:36:32.649980 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:32 crc kubenswrapper[4809]: I1127 17:36:32.696996 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" podStartSLOduration=24.696961993 podStartE2EDuration="24.696961993s" podCreationTimestamp="2025-11-27 17:36:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:36:32.686535889 +0000 UTC m=+1627.958993251" watchObservedRunningTime="2025-11-27 17:36:32.696961993 +0000 UTC m=+1627.969419345" Nov 27 17:36:33 crc kubenswrapper[4809]: E1127 17:36:33.611328 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" podUID="63be6575-532b-487f-97a2-d9fd077e5be0" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.663651 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" event={"ID":"6016f85d-a0f4-45ee-a96f-95dcb2c57e3c","Type":"ContainerStarted","Data":"e91cca17d2f674d4db2b5e8e86687956c3f6dd43a24811fd148fe4daec190eab"} Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.692917 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" event={"ID":"63be6575-532b-487f-97a2-d9fd077e5be0","Type":"ContainerStarted","Data":"ee32580d410351da0c6137c7a5e81158b87b6fe0c6595196f995ca475683acab"} Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.704218 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" event={"ID":"ad51c9a8-6543-403b-b8de-ff189da761a4","Type":"ContainerStarted","Data":"08f3f7da8331577f4f8804d44155022c53956a4953f14dde1ccdc46cf6335f02"} Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.704902 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.711999 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.726792 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" event={"ID":"976fdd57-7e54-4896-abca-914b86a98b19","Type":"ContainerStarted","Data":"25d3b6d5e5916d4e1b9a7100f92c5a7451a63274f52df859c4a1e3a0aa2ad7ef"} Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.728681 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.732510 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.739034 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" event={"ID":"5d49d329-6737-4bab-b243-64e618c0fa3b","Type":"ContainerStarted","Data":"bc1a925a46c44eb18a301814e542a26e679bfa4a28e82aa7a45eaf75a9e520e0"} Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.759047 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" event={"ID":"4e5fbc76-cd41-432e-8c5b-90ff77d0704c","Type":"ContainerStarted","Data":"89a2c66cec742b5c0576f33843bb512a0c2a6e449af1e576aa5e01a96168d4a0"} Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.760176 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.766375 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.775154 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-bzjd8" podStartSLOduration=3.42613225 podStartE2EDuration="26.775116338s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.510097018 +0000 UTC m=+1604.782554370" lastFinishedPulling="2025-11-27 17:36:32.859081106 +0000 UTC m=+1628.131538458" observedRunningTime="2025-11-27 17:36:33.755133325 +0000 UTC m=+1629.027590677" watchObservedRunningTime="2025-11-27 17:36:33.775116338 +0000 UTC m=+1629.047573690" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.796780 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-x9qq5" podStartSLOduration=3.165719282 podStartE2EDuration="26.796765097s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.261675826 +0000 UTC m=+1604.534133178" lastFinishedPulling="2025-11-27 17:36:32.892721651 +0000 UTC m=+1628.165178993" observedRunningTime="2025-11-27 17:36:33.795623306 +0000 UTC m=+1629.068080658" watchObservedRunningTime="2025-11-27 17:36:33.796765097 +0000 UTC m=+1629.069222449" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.813788 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" event={"ID":"c077fc08-1ee2-493e-b631-f01ccd0b7c6e","Type":"ContainerStarted","Data":"35cc32e5a8fbe3eff6461740fbd1f037b65ebce750b3e4f3d2b0ecb5d6860ea8"} Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.851840 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" event={"ID":"8798a938-aa20-40e2-be26-408026fb123e","Type":"ContainerStarted","Data":"ef106696dbb4f7b11ef351ea42c3750542fbb500fd7c24718a1223f170dfd0f9"} Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.842716 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-g7pm5" podStartSLOduration=2.530596186 podStartE2EDuration="26.842695958s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:08.614952963 +0000 UTC m=+1603.887410305" lastFinishedPulling="2025-11-27 17:36:32.927052725 +0000 UTC m=+1628.199510077" observedRunningTime="2025-11-27 17:36:33.841522185 +0000 UTC m=+1629.113979537" watchObservedRunningTime="2025-11-27 17:36:33.842695958 +0000 UTC m=+1629.115153310" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.876843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" event={"ID":"be316447-6b83-4563-b75e-d4088882b695","Type":"ContainerStarted","Data":"e5f11bd0c3cf82f2b310df041a113bff7fbf2936a1bc89914ae39884b1df0330"} Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.877374 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.888345 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" Nov 27 17:36:33 crc kubenswrapper[4809]: I1127 17:36:33.911511 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-swx59" podStartSLOduration=2.632096939 podStartE2EDuration="26.91149139s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:08.7441318 +0000 UTC m=+1604.016589142" lastFinishedPulling="2025-11-27 17:36:33.023526241 +0000 UTC m=+1628.295983593" observedRunningTime="2025-11-27 17:36:33.907945043 +0000 UTC m=+1629.180402415" watchObservedRunningTime="2025-11-27 17:36:33.91149139 +0000 UTC m=+1629.183948742" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.885025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" event={"ID":"6016f85d-a0f4-45ee-a96f-95dcb2c57e3c","Type":"ContainerStarted","Data":"647ecd10e1f6233caeb19490a529862f31d095e28d7094f9c646932be382088b"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.885623 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.887493 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" event={"ID":"db1b7072-b244-4e32-ba19-b8cd4254af2f","Type":"ContainerStarted","Data":"4ec5ed72398b4c579963ed8162fec03725f01bf169db6248116f71c73f817f4d"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.887721 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.889650 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" event={"ID":"8ace4cdc-1253-4a0f-b667-83e997947f9b","Type":"ContainerStarted","Data":"61ed2e0ebd14b717197d7864ab5f89553e4b6302a03697523f90dea2439db0bc"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.889941 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.891121 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.892088 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" event={"ID":"f6774dea-1e4a-4cd0-8688-5a588fe6d49b","Type":"ContainerStarted","Data":"a0d8dc66d9400c3f6bf97890eb296ba64af805797f4bb1130c80280287797dec"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.892902 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.893259 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.895021 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.895152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" event={"ID":"63be6575-532b-487f-97a2-d9fd077e5be0","Type":"ContainerStarted","Data":"b0fa3edfedb952c6549d6118b669b9ddcd3ce63d9f21ab04a8d3e70f42d07fe7"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.895699 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.897880 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" event={"ID":"27f6291c-27aa-430f-864e-05be868c89b9","Type":"ContainerStarted","Data":"a0398c3220534ca5710a3d788bc2be668bb09acd3d46c0237a8f1d8785e582e0"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.899595 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.900055 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.902001 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" event={"ID":"c077fc08-1ee2-493e-b631-f01ccd0b7c6e","Type":"ContainerStarted","Data":"cb4025494b71b97b8168495c7c9db1b122f14b48946480adaa0d5e13596d24ae"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.902126 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.904093 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" event={"ID":"8f8e02f4-c34c-4b0a-b18f-8e088d48e12c","Type":"ContainerStarted","Data":"210d6a8ac7e1cbc9893d142505e535208a1d5836ea40c845ac04d5ecc5d7b789"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.904839 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.910396 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.910880 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" event={"ID":"9fb9364a-e568-418b-b403-dfa071e60297","Type":"ContainerStarted","Data":"e00a9e198d99a259b0f895b117f69a913f3a3e6fc1faa74aa62d36217ec79271"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.912064 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.914254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" event={"ID":"8798a938-aa20-40e2-be26-408026fb123e","Type":"ContainerStarted","Data":"570a024a11758269c6f2e44702c365f5f35472496e909e320d6e7eb9fe073623"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.914511 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.925171 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" event={"ID":"5d49d329-6737-4bab-b243-64e618c0fa3b","Type":"ContainerStarted","Data":"3fdc762cbcc39a361a71213d35d56048b24169c185e711a132fa31e8b3a8c1f9"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.925240 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.925265 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.925087 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" podStartSLOduration=5.954000763 podStartE2EDuration="27.925063307s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.738173525 +0000 UTC m=+1605.010630877" lastFinishedPulling="2025-11-27 17:36:31.709236079 +0000 UTC m=+1626.981693421" observedRunningTime="2025-11-27 17:36:34.914761987 +0000 UTC m=+1630.187219359" watchObservedRunningTime="2025-11-27 17:36:34.925063307 +0000 UTC m=+1630.197520659" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.927001 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" event={"ID":"3da94235-386c-4f05-b869-2e08cee40d5d","Type":"ContainerStarted","Data":"7e1d6d134ab562d607b608c16b8c9cc728c36792409380a3b832b415c9652b47"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.927033 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" event={"ID":"3da94235-386c-4f05-b869-2e08cee40d5d","Type":"ContainerStarted","Data":"a3e5d5cc5a5196c3f75e7a7ae411e95ee7df2f83fd25818113eabbb6585f81f5"} Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.927244 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.952589 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-6t8sk" podStartSLOduration=3.777371671 podStartE2EDuration="27.952567876s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.122956801 +0000 UTC m=+1604.395414153" lastFinishedPulling="2025-11-27 17:36:33.298153006 +0000 UTC m=+1628.570610358" observedRunningTime="2025-11-27 17:36:34.948384723 +0000 UTC m=+1630.220842085" watchObservedRunningTime="2025-11-27 17:36:34.952567876 +0000 UTC m=+1630.225025228" Nov 27 17:36:34 crc kubenswrapper[4809]: I1127 17:36:34.971549 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-x4fdl" podStartSLOduration=3.958627365 podStartE2EDuration="27.971529972s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:08.965510146 +0000 UTC m=+1604.237967498" lastFinishedPulling="2025-11-27 17:36:32.978412753 +0000 UTC m=+1628.250870105" observedRunningTime="2025-11-27 17:36:34.970538205 +0000 UTC m=+1630.242995557" watchObservedRunningTime="2025-11-27 17:36:34.971529972 +0000 UTC m=+1630.243987324" Nov 27 17:36:35 crc kubenswrapper[4809]: I1127 17:36:35.006721 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-4rlvc" podStartSLOduration=4.079055661 podStartE2EDuration="28.006698939s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.112909497 +0000 UTC m=+1604.385366849" lastFinishedPulling="2025-11-27 17:36:33.040552775 +0000 UTC m=+1628.313010127" observedRunningTime="2025-11-27 17:36:34.991467835 +0000 UTC m=+1630.263925187" watchObservedRunningTime="2025-11-27 17:36:35.006698939 +0000 UTC m=+1630.279156291" Nov 27 17:36:35 crc kubenswrapper[4809]: I1127 17:36:35.045863 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-tcwh6" podStartSLOduration=3.785779749 podStartE2EDuration="28.045829874s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.505195664 +0000 UTC m=+1604.777653016" lastFinishedPulling="2025-11-27 17:36:33.765245789 +0000 UTC m=+1629.037703141" observedRunningTime="2025-11-27 17:36:35.028496333 +0000 UTC m=+1630.300953685" watchObservedRunningTime="2025-11-27 17:36:35.045829874 +0000 UTC m=+1630.318287226" Nov 27 17:36:35 crc kubenswrapper[4809]: I1127 17:36:35.075836 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-c9pds" podStartSLOduration=4.53275774 podStartE2EDuration="28.07581375s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.655045113 +0000 UTC m=+1604.927502465" lastFinishedPulling="2025-11-27 17:36:33.198101123 +0000 UTC m=+1628.470558475" observedRunningTime="2025-11-27 17:36:35.075511592 +0000 UTC m=+1630.347968944" watchObservedRunningTime="2025-11-27 17:36:35.07581375 +0000 UTC m=+1630.348271102" Nov 27 17:36:35 crc kubenswrapper[4809]: I1127 17:36:35.156895 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" podStartSLOduration=6.453379884 podStartE2EDuration="27.156872697s" podCreationTimestamp="2025-11-27 17:36:08 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.768237264 +0000 UTC m=+1605.040694616" lastFinishedPulling="2025-11-27 17:36:30.471730077 +0000 UTC m=+1625.744187429" observedRunningTime="2025-11-27 17:36:35.126080608 +0000 UTC m=+1630.398537960" watchObservedRunningTime="2025-11-27 17:36:35.156872697 +0000 UTC m=+1630.429330049" Nov 27 17:36:35 crc kubenswrapper[4809]: I1127 17:36:35.210540 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" podStartSLOduration=3.373685163 podStartE2EDuration="28.210509137s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.355899331 +0000 UTC m=+1604.628356683" lastFinishedPulling="2025-11-27 17:36:34.192723305 +0000 UTC m=+1629.465180657" observedRunningTime="2025-11-27 17:36:35.15554518 +0000 UTC m=+1630.428002542" watchObservedRunningTime="2025-11-27 17:36:35.210509137 +0000 UTC m=+1630.482966489" Nov 27 17:36:35 crc kubenswrapper[4809]: I1127 17:36:35.251643 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" podStartSLOduration=21.768846758 podStartE2EDuration="28.251617256s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:25.423906245 +0000 UTC m=+1620.696363597" lastFinishedPulling="2025-11-27 17:36:31.906676743 +0000 UTC m=+1627.179134095" observedRunningTime="2025-11-27 17:36:35.207195997 +0000 UTC m=+1630.479653349" watchObservedRunningTime="2025-11-27 17:36:35.251617256 +0000 UTC m=+1630.524074608" Nov 27 17:36:35 crc kubenswrapper[4809]: I1127 17:36:35.252721 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d77b94747-vvtmf" podStartSLOduration=4.824826261 podStartE2EDuration="28.252712706s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.638159194 +0000 UTC m=+1604.910616546" lastFinishedPulling="2025-11-27 17:36:33.066045639 +0000 UTC m=+1628.338502991" observedRunningTime="2025-11-27 17:36:35.246294401 +0000 UTC m=+1630.518751763" watchObservedRunningTime="2025-11-27 17:36:35.252712706 +0000 UTC m=+1630.525170068" Nov 27 17:36:35 crc kubenswrapper[4809]: I1127 17:36:35.286562 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" podStartSLOduration=5.120449944 podStartE2EDuration="27.286543037s" podCreationTimestamp="2025-11-27 17:36:08 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.740489048 +0000 UTC m=+1605.012946400" lastFinishedPulling="2025-11-27 17:36:31.90658214 +0000 UTC m=+1627.179039493" observedRunningTime="2025-11-27 17:36:35.270612054 +0000 UTC m=+1630.543069426" watchObservedRunningTime="2025-11-27 17:36:35.286543037 +0000 UTC m=+1630.559000379" Nov 27 17:36:35 crc kubenswrapper[4809]: I1127 17:36:35.350040 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" podStartSLOduration=6.078886805 podStartE2EDuration="28.350018145s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.741467996 +0000 UTC m=+1605.013925348" lastFinishedPulling="2025-11-27 17:36:32.012599336 +0000 UTC m=+1627.285056688" observedRunningTime="2025-11-27 17:36:35.34802881 +0000 UTC m=+1630.620486172" watchObservedRunningTime="2025-11-27 17:36:35.350018145 +0000 UTC m=+1630.622475497" Nov 27 17:36:37 crc kubenswrapper[4809]: I1127 17:36:37.469872 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" Nov 27 17:36:38 crc kubenswrapper[4809]: I1127 17:36:38.907618 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8bq9l" Nov 27 17:36:38 crc kubenswrapper[4809]: I1127 17:36:38.957013 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-ff79b6df5-lq72j" Nov 27 17:36:38 crc kubenswrapper[4809]: I1127 17:36:38.999725 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-j6nv4" Nov 27 17:36:39 crc kubenswrapper[4809]: I1127 17:36:39.111866 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-d5hqw" Nov 27 17:36:39 crc kubenswrapper[4809]: I1127 17:36:39.812782 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:39 crc kubenswrapper[4809]: I1127 17:36:39.820706 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1-cert\") pod \"infra-operator-controller-manager-57548d458d-b28r8\" (UID: \"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:39 crc kubenswrapper[4809]: I1127 17:36:39.996347 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vfclg" Nov 27 17:36:40 crc kubenswrapper[4809]: I1127 17:36:40.004792 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:40 crc kubenswrapper[4809]: I1127 17:36:40.447792 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-b28r8"] Nov 27 17:36:40 crc kubenswrapper[4809]: W1127 17:36:40.452948 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa9f8064_5a3b_4ed8_8c85_c77f54e22ff1.slice/crio-6013f276ff877f579ea7f71eae55ac13f346451aba12720b4bfc7df5e5df50bd WatchSource:0}: Error finding container 6013f276ff877f579ea7f71eae55ac13f346451aba12720b4bfc7df5e5df50bd: Status 404 returned error can't find the container with id 6013f276ff877f579ea7f71eae55ac13f346451aba12720b4bfc7df5e5df50bd Nov 27 17:36:40 crc kubenswrapper[4809]: I1127 17:36:40.991856 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" event={"ID":"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1","Type":"ContainerStarted","Data":"6013f276ff877f579ea7f71eae55ac13f346451aba12720b4bfc7df5e5df50bd"} Nov 27 17:36:44 crc kubenswrapper[4809]: I1127 17:36:44.016887 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" event={"ID":"b470d686-15fd-4d31-ac6a-ffb60d3c9d79","Type":"ContainerStarted","Data":"f2aaed3376531e4b5f1d80e8f87f2114df1ce251f6900cc3d9ffd961661e0d40"} Nov 27 17:36:44 crc kubenswrapper[4809]: I1127 17:36:44.018967 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" event={"ID":"1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659","Type":"ContainerStarted","Data":"158bb322efa2614b8286ffbbb753050b5294b2e7b85891045c10ba259e5b690e"} Nov 27 17:36:44 crc kubenswrapper[4809]: I1127 17:36:44.021086 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" event={"ID":"da239d06-b1c7-4b06-b830-5108030beed6","Type":"ContainerStarted","Data":"b7381d51d15b18e784f53d927bb3974ad3a10117170c545e6fb4cedd7617ab26"} Nov 27 17:36:44 crc kubenswrapper[4809]: I1127 17:36:44.036407 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-955677c94-l58tv" podStartSLOduration=22.971964363 podStartE2EDuration="37.036377723s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:08.892698504 +0000 UTC m=+1604.165155866" lastFinishedPulling="2025-11-27 17:36:22.957111874 +0000 UTC m=+1618.229569226" observedRunningTime="2025-11-27 17:36:44.030859123 +0000 UTC m=+1639.303316475" watchObservedRunningTime="2025-11-27 17:36:44.036377723 +0000 UTC m=+1639.308835105" Nov 27 17:36:44 crc kubenswrapper[4809]: I1127 17:36:44.081373 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ll5l2" podStartSLOduration=24.49187088 podStartE2EDuration="37.081351927s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.342065894 +0000 UTC m=+1604.614523246" lastFinishedPulling="2025-11-27 17:36:21.931546941 +0000 UTC m=+1617.204004293" observedRunningTime="2025-11-27 17:36:44.051287638 +0000 UTC m=+1639.323745000" watchObservedRunningTime="2025-11-27 17:36:44.081351927 +0000 UTC m=+1639.353809279" Nov 27 17:36:44 crc kubenswrapper[4809]: I1127 17:36:44.278439 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm" Nov 27 17:36:44 crc kubenswrapper[4809]: I1127 17:36:44.860614 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6644d5b8df-l9kv7" Nov 27 17:36:45 crc kubenswrapper[4809]: I1127 17:36:45.046751 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-xrjzd" podStartSLOduration=24.624250504 podStartE2EDuration="38.046711312s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.487157714 +0000 UTC m=+1604.759615066" lastFinishedPulling="2025-11-27 17:36:22.909618522 +0000 UTC m=+1618.182075874" observedRunningTime="2025-11-27 17:36:45.04151936 +0000 UTC m=+1640.313976722" watchObservedRunningTime="2025-11-27 17:36:45.046711312 +0000 UTC m=+1640.319168664" Nov 27 17:36:46 crc kubenswrapper[4809]: I1127 17:36:46.037900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" event={"ID":"16fc4440-2960-4b7d-b059-c5192ea37a50","Type":"ContainerStarted","Data":"73bcbf3f5bdabe07448212183a4a42ec2b0345c7b36e90bb89418e0981f11907"} Nov 27 17:36:47 crc kubenswrapper[4809]: I1127 17:36:47.073658 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-mkngh" podStartSLOduration=26.226384059 podStartE2EDuration="40.069543809s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:09.11817304 +0000 UTC m=+1604.390630392" lastFinishedPulling="2025-11-27 17:36:22.96133279 +0000 UTC m=+1618.233790142" observedRunningTime="2025-11-27 17:36:47.057508582 +0000 UTC m=+1642.329965934" watchObservedRunningTime="2025-11-27 17:36:47.069543809 +0000 UTC m=+1642.342001161" Nov 27 17:36:48 crc kubenswrapper[4809]: I1127 17:36:48.376161 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-mwv49" Nov 27 17:36:49 crc kubenswrapper[4809]: I1127 17:36:49.057942 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" event={"ID":"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1","Type":"ContainerStarted","Data":"35d8405418ee65984b6922cddb2f44dcb2ad3cdf322712bc84f27d5f1af582f4"} Nov 27 17:36:49 crc kubenswrapper[4809]: I1127 17:36:49.057997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" event={"ID":"aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1","Type":"ContainerStarted","Data":"039cd1c02588ae70b825d236826feff172d38be8b60651b5ff9a38c62bbc9c03"} Nov 27 17:36:49 crc kubenswrapper[4809]: I1127 17:36:49.058121 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:36:49 crc kubenswrapper[4809]: I1127 17:36:49.073523 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" podStartSLOduration=34.489688966 podStartE2EDuration="42.073504463s" podCreationTimestamp="2025-11-27 17:36:07 +0000 UTC" firstStartedPulling="2025-11-27 17:36:40.455161008 +0000 UTC m=+1635.727618360" lastFinishedPulling="2025-11-27 17:36:48.038976505 +0000 UTC m=+1643.311433857" observedRunningTime="2025-11-27 17:36:49.071774386 +0000 UTC m=+1644.344231738" watchObservedRunningTime="2025-11-27 17:36:49.073504463 +0000 UTC m=+1644.345961815" Nov 27 17:36:55 crc kubenswrapper[4809]: I1127 17:36:55.779982 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:36:55 crc kubenswrapper[4809]: I1127 17:36:55.780455 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:37:00 crc kubenswrapper[4809]: I1127 17:37:00.010881 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-b28r8" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.154697 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9cxbm"] Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.160082 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.163964 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9cxbm"] Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.170335 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-srx9b" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.170509 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.170765 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.170897 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.240515 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qfcd2"] Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.242543 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.251113 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.269911 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qfcd2"] Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.287313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-qfcd2\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.287363 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-config\") pod \"dnsmasq-dns-78dd6ddcc-qfcd2\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.287388 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jxvw\" (UniqueName: \"kubernetes.io/projected/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-kube-api-access-9jxvw\") pod \"dnsmasq-dns-675f4bcbfc-9cxbm\" (UID: \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.287458 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrn2m\" (UniqueName: \"kubernetes.io/projected/d336542d-4cea-4b20-ae0b-c27ba218d364-kube-api-access-rrn2m\") pod \"dnsmasq-dns-78dd6ddcc-qfcd2\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.287481 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-config\") pod \"dnsmasq-dns-675f4bcbfc-9cxbm\" (UID: \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.388557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrn2m\" (UniqueName: \"kubernetes.io/projected/d336542d-4cea-4b20-ae0b-c27ba218d364-kube-api-access-rrn2m\") pod \"dnsmasq-dns-78dd6ddcc-qfcd2\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.388604 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-config\") pod \"dnsmasq-dns-675f4bcbfc-9cxbm\" (UID: \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.388637 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-qfcd2\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.388665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-config\") pod \"dnsmasq-dns-78dd6ddcc-qfcd2\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.388690 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jxvw\" (UniqueName: \"kubernetes.io/projected/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-kube-api-access-9jxvw\") pod \"dnsmasq-dns-675f4bcbfc-9cxbm\" (UID: \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.389507 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-config\") pod \"dnsmasq-dns-675f4bcbfc-9cxbm\" (UID: \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.389516 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-qfcd2\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.389570 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-config\") pod \"dnsmasq-dns-78dd6ddcc-qfcd2\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.408230 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrn2m\" (UniqueName: \"kubernetes.io/projected/d336542d-4cea-4b20-ae0b-c27ba218d364-kube-api-access-rrn2m\") pod \"dnsmasq-dns-78dd6ddcc-qfcd2\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.408260 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jxvw\" (UniqueName: \"kubernetes.io/projected/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-kube-api-access-9jxvw\") pod \"dnsmasq-dns-675f4bcbfc-9cxbm\" (UID: \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.496730 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.621286 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:16 crc kubenswrapper[4809]: I1127 17:37:16.925089 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9cxbm"] Nov 27 17:37:17 crc kubenswrapper[4809]: I1127 17:37:17.051250 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qfcd2"] Nov 27 17:37:17 crc kubenswrapper[4809]: W1127 17:37:17.052276 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd336542d_4cea_4b20_ae0b_c27ba218d364.slice/crio-a126692fe7704bb582ef820bb9506afe0214004efac9ca1b26de4410c50be828 WatchSource:0}: Error finding container a126692fe7704bb582ef820bb9506afe0214004efac9ca1b26de4410c50be828: Status 404 returned error can't find the container with id a126692fe7704bb582ef820bb9506afe0214004efac9ca1b26de4410c50be828 Nov 27 17:37:17 crc kubenswrapper[4809]: I1127 17:37:17.265256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" event={"ID":"d336542d-4cea-4b20-ae0b-c27ba218d364","Type":"ContainerStarted","Data":"a126692fe7704bb582ef820bb9506afe0214004efac9ca1b26de4410c50be828"} Nov 27 17:37:17 crc kubenswrapper[4809]: I1127 17:37:17.266348 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" event={"ID":"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f","Type":"ContainerStarted","Data":"da909169d9a2d690291c997e3e52932245d30a10cb8183fbd53b2cca390ce1f4"} Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.003241 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9cxbm"] Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.038986 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-2qrtg"] Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.040677 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.058944 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-2qrtg"] Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.241217 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-dns-svc\") pod \"dnsmasq-dns-666b6646f7-2qrtg\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.241313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj24w\" (UniqueName: \"kubernetes.io/projected/a67727fd-a1c7-4d34-897c-9522dd801a83-kube-api-access-nj24w\") pod \"dnsmasq-dns-666b6646f7-2qrtg\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.241382 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-config\") pod \"dnsmasq-dns-666b6646f7-2qrtg\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.341988 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qfcd2"] Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.342602 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-dns-svc\") pod \"dnsmasq-dns-666b6646f7-2qrtg\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.342678 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj24w\" (UniqueName: \"kubernetes.io/projected/a67727fd-a1c7-4d34-897c-9522dd801a83-kube-api-access-nj24w\") pod \"dnsmasq-dns-666b6646f7-2qrtg\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.342717 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-config\") pod \"dnsmasq-dns-666b6646f7-2qrtg\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.343669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-config\") pod \"dnsmasq-dns-666b6646f7-2qrtg\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.343670 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-dns-svc\") pod \"dnsmasq-dns-666b6646f7-2qrtg\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.365595 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pmp8v"] Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.380355 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pmp8v"] Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.380455 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.380546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj24w\" (UniqueName: \"kubernetes.io/projected/a67727fd-a1c7-4d34-897c-9522dd801a83-kube-api-access-nj24w\") pod \"dnsmasq-dns-666b6646f7-2qrtg\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.546263 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-config\") pod \"dnsmasq-dns-57d769cc4f-pmp8v\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.546389 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf5zg\" (UniqueName: \"kubernetes.io/projected/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-kube-api-access-sf5zg\") pod \"dnsmasq-dns-57d769cc4f-pmp8v\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.546417 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-pmp8v\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.647942 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf5zg\" (UniqueName: \"kubernetes.io/projected/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-kube-api-access-sf5zg\") pod \"dnsmasq-dns-57d769cc4f-pmp8v\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.647998 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-pmp8v\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.648060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-config\") pod \"dnsmasq-dns-57d769cc4f-pmp8v\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.649606 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-config\") pod \"dnsmasq-dns-57d769cc4f-pmp8v\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.650066 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-pmp8v\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.674471 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.688858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf5zg\" (UniqueName: \"kubernetes.io/projected/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-kube-api-access-sf5zg\") pod \"dnsmasq-dns-57d769cc4f-pmp8v\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:19 crc kubenswrapper[4809]: I1127 17:37:19.749039 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.170848 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.172269 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.175027 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.175096 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5mnsp" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.175181 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.176031 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.176245 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.176852 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-2qrtg"] Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.177155 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.177171 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.189543 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259028 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259085 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259125 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259149 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259180 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-config-data\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259213 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259241 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01288244-f670-47a4-ac43-180191d7f331-pod-info\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259265 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01288244-f670-47a4-ac43-180191d7f331-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-server-conf\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259370 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz6n6\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-kube-api-access-nz6n6\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.259417 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.263304 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pmp8v"] Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.308817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" event={"ID":"a67727fd-a1c7-4d34-897c-9522dd801a83","Type":"ContainerStarted","Data":"bbba2158a914c552a7466307b357dec59073058db0fa2c078ef0a4a88eb9fae0"} Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361240 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz6n6\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-kube-api-access-nz6n6\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361360 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361384 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361425 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361448 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361479 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-config-data\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361511 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361543 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01288244-f670-47a4-ac43-180191d7f331-pod-info\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361568 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01288244-f670-47a4-ac43-180191d7f331-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.361613 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-server-conf\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.362645 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.362991 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.363011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-server-conf\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.363348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.363607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.363891 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-config-data\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.368015 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.368207 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01288244-f670-47a4-ac43-180191d7f331-pod-info\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.369203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01288244-f670-47a4-ac43-180191d7f331-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.369544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.381168 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz6n6\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-kube-api-access-nz6n6\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.387223 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.483978 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.485509 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.489893 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.489910 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.491464 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.493980 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.494326 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.494504 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.494771 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6rbpw" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.496090 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.501170 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.565485 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8145e23-a84e-405f-beb6-e27bb2aa1c59-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.565537 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdlbl\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-kube-api-access-sdlbl\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.565567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.565605 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8145e23-a84e-405f-beb6-e27bb2aa1c59-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.565657 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.565685 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.565706 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.565812 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.565897 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.566017 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.566146 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668449 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668519 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668571 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668598 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668649 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668708 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8145e23-a84e-405f-beb6-e27bb2aa1c59-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668871 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdlbl\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-kube-api-access-sdlbl\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668895 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.668930 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8145e23-a84e-405f-beb6-e27bb2aa1c59-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.669364 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.669685 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.669719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.670291 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.670869 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.671092 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.673504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.674284 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.675104 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8145e23-a84e-405f-beb6-e27bb2aa1c59-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.683422 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8145e23-a84e-405f-beb6-e27bb2aa1c59-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.689099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdlbl\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-kube-api-access-sdlbl\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.721094 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:20 crc kubenswrapper[4809]: I1127 17:37:20.819107 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.186535 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.188334 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.192843 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.193442 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.193882 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.194006 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-hntxx" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.199711 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.221325 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.295035 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.295105 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.295128 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-996kp\" (UniqueName: \"kubernetes.io/projected/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-kube-api-access-996kp\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.295158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-config-data-default\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.295268 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.295301 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-kolla-config\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.295381 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.295411 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.396822 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-config-data-default\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.397123 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.397211 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-kolla-config\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.397333 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.397408 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.397355 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.397602 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.398430 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.398546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-996kp\" (UniqueName: \"kubernetes.io/projected/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-kube-api-access-996kp\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.398621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.397955 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.398134 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-kolla-config\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.409066 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.418322 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-config-data-default\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.419575 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.427323 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.447452 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-996kp\" (UniqueName: \"kubernetes.io/projected/b3cf12fa-2068-4bcd-ae42-d7c6dfea6782-kube-api-access-996kp\") pod \"openstack-galera-0\" (UID: \"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782\") " pod="openstack/openstack-galera-0" Nov 27 17:37:22 crc kubenswrapper[4809]: I1127 17:37:22.529422 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.343170 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" event={"ID":"c9e56ebf-c6e1-4675-afcb-cbb600ac778e","Type":"ContainerStarted","Data":"7823488e5020c66057c05bc12141c4300559609ef470f7313a1fc0ed5b07e550"} Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.692677 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.694210 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.703962 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.704079 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8m2t4" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.704129 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.704583 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.718762 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.823249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f82a4054-fea3-4071-b65d-529d900c0f04-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.823569 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f82a4054-fea3-4071-b65d-529d900c0f04-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.823759 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f82a4054-fea3-4071-b65d-529d900c0f04-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.823881 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f82a4054-fea3-4071-b65d-529d900c0f04-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.824005 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.824283 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpzv5\" (UniqueName: \"kubernetes.io/projected/f82a4054-fea3-4071-b65d-529d900c0f04-kube-api-access-fpzv5\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.824437 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f82a4054-fea3-4071-b65d-529d900c0f04-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.824565 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f82a4054-fea3-4071-b65d-529d900c0f04-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.858425 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.859906 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.862433 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-b8gdh" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.862727 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.863385 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.887500 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.926916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f82a4054-fea3-4071-b65d-529d900c0f04-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.926975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb517228-deb6-41fa-b6aa-354014cd34e1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927015 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f82a4054-fea3-4071-b65d-529d900c0f04-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927058 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb517228-deb6-41fa-b6aa-354014cd34e1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927099 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f82a4054-fea3-4071-b65d-529d900c0f04-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927138 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb4db\" (UniqueName: \"kubernetes.io/projected/bb517228-deb6-41fa-b6aa-354014cd34e1-kube-api-access-nb4db\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f82a4054-fea3-4071-b65d-529d900c0f04-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927225 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f82a4054-fea3-4071-b65d-529d900c0f04-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f82a4054-fea3-4071-b65d-529d900c0f04-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb517228-deb6-41fa-b6aa-354014cd34e1-kolla-config\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927325 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927352 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb517228-deb6-41fa-b6aa-354014cd34e1-config-data\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927378 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f82a4054-fea3-4071-b65d-529d900c0f04-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.927393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpzv5\" (UniqueName: \"kubernetes.io/projected/f82a4054-fea3-4071-b65d-529d900c0f04-kube-api-access-fpzv5\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.928049 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f82a4054-fea3-4071-b65d-529d900c0f04-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.928350 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.928363 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f82a4054-fea3-4071-b65d-529d900c0f04-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.929812 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f82a4054-fea3-4071-b65d-529d900c0f04-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.933701 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f82a4054-fea3-4071-b65d-529d900c0f04-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.952437 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpzv5\" (UniqueName: \"kubernetes.io/projected/f82a4054-fea3-4071-b65d-529d900c0f04-kube-api-access-fpzv5\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:23 crc kubenswrapper[4809]: I1127 17:37:23.952827 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f82a4054-fea3-4071-b65d-529d900c0f04-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.012942 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f82a4054-fea3-4071-b65d-529d900c0f04\") " pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.031017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb4db\" (UniqueName: \"kubernetes.io/projected/bb517228-deb6-41fa-b6aa-354014cd34e1-kube-api-access-nb4db\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.031196 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb517228-deb6-41fa-b6aa-354014cd34e1-kolla-config\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.031292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb517228-deb6-41fa-b6aa-354014cd34e1-config-data\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.031402 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb517228-deb6-41fa-b6aa-354014cd34e1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.031488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb517228-deb6-41fa-b6aa-354014cd34e1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.032937 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb517228-deb6-41fa-b6aa-354014cd34e1-kolla-config\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.039436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb517228-deb6-41fa-b6aa-354014cd34e1-config-data\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.042405 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb517228-deb6-41fa-b6aa-354014cd34e1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.048338 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb517228-deb6-41fa-b6aa-354014cd34e1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.063097 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb4db\" (UniqueName: \"kubernetes.io/projected/bb517228-deb6-41fa-b6aa-354014cd34e1-kube-api-access-nb4db\") pod \"memcached-0\" (UID: \"bb517228-deb6-41fa-b6aa-354014cd34e1\") " pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.192571 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 27 17:37:24 crc kubenswrapper[4809]: I1127 17:37:24.316223 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.779383 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.779764 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.779838 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.780610 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.780669 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" gracePeriod=600 Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.809890 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.810969 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.813167 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-k2lgt" Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.822989 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.862688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w57f\" (UniqueName: \"kubernetes.io/projected/07137204-40d7-4004-a77d-6b704e1d1602-kube-api-access-6w57f\") pod \"kube-state-metrics-0\" (UID: \"07137204-40d7-4004-a77d-6b704e1d1602\") " pod="openstack/kube-state-metrics-0" Nov 27 17:37:25 crc kubenswrapper[4809]: I1127 17:37:25.964077 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w57f\" (UniqueName: \"kubernetes.io/projected/07137204-40d7-4004-a77d-6b704e1d1602-kube-api-access-6w57f\") pod \"kube-state-metrics-0\" (UID: \"07137204-40d7-4004-a77d-6b704e1d1602\") " pod="openstack/kube-state-metrics-0" Nov 27 17:37:26 crc kubenswrapper[4809]: I1127 17:37:26.000659 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w57f\" (UniqueName: \"kubernetes.io/projected/07137204-40d7-4004-a77d-6b704e1d1602-kube-api-access-6w57f\") pod \"kube-state-metrics-0\" (UID: \"07137204-40d7-4004-a77d-6b704e1d1602\") " pod="openstack/kube-state-metrics-0" Nov 27 17:37:26 crc kubenswrapper[4809]: I1127 17:37:26.142041 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 17:37:26 crc kubenswrapper[4809]: I1127 17:37:26.382537 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" exitCode=0 Nov 27 17:37:26 crc kubenswrapper[4809]: I1127 17:37:26.382591 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596"} Nov 27 17:37:26 crc kubenswrapper[4809]: I1127 17:37:26.382634 4809 scope.go:117] "RemoveContainer" containerID="de4c3ebc223ee9f9580cd86c386eb6d52d4734548883eb67a85381672ea09fa4" Nov 27 17:37:29 crc kubenswrapper[4809]: I1127 17:37:29.865931 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-h8g6t"] Nov 27 17:37:29 crc kubenswrapper[4809]: I1127 17:37:29.867467 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:29 crc kubenswrapper[4809]: I1127 17:37:29.872866 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-bgnbw" Nov 27 17:37:29 crc kubenswrapper[4809]: I1127 17:37:29.872922 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 27 17:37:29 crc kubenswrapper[4809]: I1127 17:37:29.873149 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 27 17:37:29 crc kubenswrapper[4809]: I1127 17:37:29.879629 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h8g6t"] Nov 27 17:37:29 crc kubenswrapper[4809]: I1127 17:37:29.902706 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-jnvf9"] Nov 27 17:37:29 crc kubenswrapper[4809]: I1127 17:37:29.904412 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:29 crc kubenswrapper[4809]: I1127 17:37:29.914050 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jnvf9"] Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.030460 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/55d50ebb-8cab-42df-96b5-9598262337a4-var-run\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.030515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/55d50ebb-8cab-42df-96b5-9598262337a4-ovn-controller-tls-certs\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.030547 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55d50ebb-8cab-42df-96b5-9598262337a4-scripts\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.030571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d50ebb-8cab-42df-96b5-9598262337a4-combined-ca-bundle\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.030651 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/55d50ebb-8cab-42df-96b5-9598262337a4-var-log-ovn\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.030771 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-var-run\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.030803 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-etc-ovs\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.030819 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-var-lib\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.030841 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a501e4bf-f184-451f-9b21-a813f147413d-scripts\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.031005 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-var-log\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.031066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfx54\" (UniqueName: \"kubernetes.io/projected/a501e4bf-f184-451f-9b21-a813f147413d-kube-api-access-lfx54\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.031094 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sstn\" (UniqueName: \"kubernetes.io/projected/55d50ebb-8cab-42df-96b5-9598262337a4-kube-api-access-6sstn\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.031154 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/55d50ebb-8cab-42df-96b5-9598262337a4-var-run-ovn\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132595 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/55d50ebb-8cab-42df-96b5-9598262337a4-ovn-controller-tls-certs\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132664 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55d50ebb-8cab-42df-96b5-9598262337a4-scripts\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132693 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d50ebb-8cab-42df-96b5-9598262337a4-combined-ca-bundle\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132730 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/55d50ebb-8cab-42df-96b5-9598262337a4-var-log-ovn\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132796 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-var-run\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-etc-ovs\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132849 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-var-lib\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132880 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a501e4bf-f184-451f-9b21-a813f147413d-scripts\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132924 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-var-log\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132960 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfx54\" (UniqueName: \"kubernetes.io/projected/a501e4bf-f184-451f-9b21-a813f147413d-kube-api-access-lfx54\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.132980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sstn\" (UniqueName: \"kubernetes.io/projected/55d50ebb-8cab-42df-96b5-9598262337a4-kube-api-access-6sstn\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.133013 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/55d50ebb-8cab-42df-96b5-9598262337a4-var-run-ovn\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.133045 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/55d50ebb-8cab-42df-96b5-9598262337a4-var-run\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.133619 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/55d50ebb-8cab-42df-96b5-9598262337a4-var-run\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.134120 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-var-lib\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.134552 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/55d50ebb-8cab-42df-96b5-9598262337a4-var-log-ovn\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.136253 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-var-log\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.136341 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-var-run\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.136655 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a501e4bf-f184-451f-9b21-a813f147413d-etc-ovs\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.136767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55d50ebb-8cab-42df-96b5-9598262337a4-scripts\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.136878 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/55d50ebb-8cab-42df-96b5-9598262337a4-var-run-ovn\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.137932 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a501e4bf-f184-451f-9b21-a813f147413d-scripts\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.146542 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d50ebb-8cab-42df-96b5-9598262337a4-combined-ca-bundle\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.146710 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/55d50ebb-8cab-42df-96b5-9598262337a4-ovn-controller-tls-certs\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.159113 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sstn\" (UniqueName: \"kubernetes.io/projected/55d50ebb-8cab-42df-96b5-9598262337a4-kube-api-access-6sstn\") pod \"ovn-controller-h8g6t\" (UID: \"55d50ebb-8cab-42df-96b5-9598262337a4\") " pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.161110 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfx54\" (UniqueName: \"kubernetes.io/projected/a501e4bf-f184-451f-9b21-a813f147413d-kube-api-access-lfx54\") pod \"ovn-controller-ovs-jnvf9\" (UID: \"a501e4bf-f184-451f-9b21-a813f147413d\") " pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.186270 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:30 crc kubenswrapper[4809]: I1127 17:37:30.224092 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:31 crc kubenswrapper[4809]: E1127 17:37:31.211977 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:37:31 crc kubenswrapper[4809]: I1127 17:37:31.438753 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:37:31 crc kubenswrapper[4809]: E1127 17:37:31.438984 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:37:32 crc kubenswrapper[4809]: E1127 17:37:32.307578 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 27 17:37:32 crc kubenswrapper[4809]: E1127 17:37:32.307730 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9jxvw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-9cxbm_openstack(b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 17:37:32 crc kubenswrapper[4809]: E1127 17:37:32.308912 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" podUID="b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f" Nov 27 17:37:32 crc kubenswrapper[4809]: E1127 17:37:32.321385 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 27 17:37:32 crc kubenswrapper[4809]: E1127 17:37:32.321560 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rrn2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-qfcd2_openstack(d336542d-4cea-4b20-ae0b-c27ba218d364): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 17:37:32 crc kubenswrapper[4809]: E1127 17:37:32.322785 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" podUID="d336542d-4cea-4b20-ae0b-c27ba218d364" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.704841 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.708791 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.712131 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.712508 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.712634 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-wktnx" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.712840 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.713076 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.720859 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.895670 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.895780 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs6kc\" (UniqueName: \"kubernetes.io/projected/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-kube-api-access-hs6kc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.895821 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-config\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.895860 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.896186 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.896269 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.896350 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.896411 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.906271 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.908428 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.912917 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.914268 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.914428 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.918322 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-9qt5c" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.925142 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998003 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs6kc\" (UniqueName: \"kubernetes.io/projected/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-kube-api-access-hs6kc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998268 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-config\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998293 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998332 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-config\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998369 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998558 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998587 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998607 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f86wh\" (UniqueName: \"kubernetes.io/projected/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-kube-api-access-f86wh\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998639 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998663 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998691 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998815 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:32 crc kubenswrapper[4809]: I1127 17:37:32.998834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:32.999685 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:32.999798 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-config\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.000844 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.004083 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.004722 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.005418 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.005459 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.014504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs6kc\" (UniqueName: \"kubernetes.io/projected/a6ab5ce4-ce26-4eee-98fe-1bb380df9376-kube-api-access-hs6kc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.036782 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ab5ce4-ce26-4eee-98fe-1bb380df9376\") " pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.080785 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.085457 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.100050 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f86wh\" (UniqueName: \"kubernetes.io/projected/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-kube-api-access-f86wh\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.100473 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.100618 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.100816 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.100961 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.101148 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-config\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.101264 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.101036 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.100477 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.102140 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.102273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.102663 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-config\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.105981 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.108481 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.110562 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.116014 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.119477 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f86wh\" (UniqueName: \"kubernetes.io/projected/cda8dbab-96fd-4d1b-a9cb-e056382df1a2-kube-api-access-f86wh\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.135785 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cda8dbab-96fd-4d1b-a9cb-e056382df1a2\") " pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.203266 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jxvw\" (UniqueName: \"kubernetes.io/projected/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-kube-api-access-9jxvw\") pod \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\" (UID: \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\") " Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.203402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-config\") pod \"d336542d-4cea-4b20-ae0b-c27ba218d364\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.203445 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-dns-svc\") pod \"d336542d-4cea-4b20-ae0b-c27ba218d364\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.203549 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-config\") pod \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\" (UID: \"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f\") " Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.203611 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrn2m\" (UniqueName: \"kubernetes.io/projected/d336542d-4cea-4b20-ae0b-c27ba218d364-kube-api-access-rrn2m\") pod \"d336542d-4cea-4b20-ae0b-c27ba218d364\" (UID: \"d336542d-4cea-4b20-ae0b-c27ba218d364\") " Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.206058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-config" (OuterVolumeSpecName: "config") pod "d336542d-4cea-4b20-ae0b-c27ba218d364" (UID: "d336542d-4cea-4b20-ae0b-c27ba218d364"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.208414 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-config" (OuterVolumeSpecName: "config") pod "b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f" (UID: "b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.208894 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d336542d-4cea-4b20-ae0b-c27ba218d364" (UID: "d336542d-4cea-4b20-ae0b-c27ba218d364"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.209497 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-kube-api-access-9jxvw" (OuterVolumeSpecName: "kube-api-access-9jxvw") pod "b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f" (UID: "b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f"). InnerVolumeSpecName "kube-api-access-9jxvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.211066 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d336542d-4cea-4b20-ae0b-c27ba218d364-kube-api-access-rrn2m" (OuterVolumeSpecName: "kube-api-access-rrn2m") pod "d336542d-4cea-4b20-ae0b-c27ba218d364" (UID: "d336542d-4cea-4b20-ae0b-c27ba218d364"). InnerVolumeSpecName "kube-api-access-rrn2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.306769 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.306837 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrn2m\" (UniqueName: \"kubernetes.io/projected/d336542d-4cea-4b20-ae0b-c27ba218d364-kube-api-access-rrn2m\") on node \"crc\" DevicePath \"\"" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.306858 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jxvw\" (UniqueName: \"kubernetes.io/projected/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f-kube-api-access-9jxvw\") on node \"crc\" DevicePath \"\"" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.306875 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.306886 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d336542d-4cea-4b20-ae0b-c27ba218d364-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.366646 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.401904 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.464647 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.525992 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.526058 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.526524 4809 generic.go:334] "Generic (PLEG): container finished" podID="c9e56ebf-c6e1-4675-afcb-cbb600ac778e" containerID="503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72" exitCode=0 Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.526715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" event={"ID":"c9e56ebf-c6e1-4675-afcb-cbb600ac778e","Type":"ContainerDied","Data":"503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72"} Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.533948 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f82a4054-fea3-4071-b65d-529d900c0f04","Type":"ContainerStarted","Data":"fcfda901e170af65191b5079b93ac2f73b9411d4153122154962c73b436bdacc"} Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.554305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" event={"ID":"b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f","Type":"ContainerDied","Data":"da909169d9a2d690291c997e3e52932245d30a10cb8183fbd53b2cca390ce1f4"} Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.554384 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9cxbm" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.558476 4809 generic.go:334] "Generic (PLEG): container finished" podID="a67727fd-a1c7-4d34-897c-9522dd801a83" containerID="af1456fc9d80c1fd1f5bae61d78fb8569bb7eac4e2d8facc31ca6c8e01789399" exitCode=0 Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.559139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" event={"ID":"a67727fd-a1c7-4d34-897c-9522dd801a83","Type":"ContainerDied","Data":"af1456fc9d80c1fd1f5bae61d78fb8569bb7eac4e2d8facc31ca6c8e01789399"} Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.568478 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" event={"ID":"d336542d-4cea-4b20-ae0b-c27ba218d364","Type":"ContainerDied","Data":"a126692fe7704bb582ef820bb9506afe0214004efac9ca1b26de4410c50be828"} Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.568574 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qfcd2" Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.656671 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qfcd2"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.675546 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qfcd2"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.685211 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jnvf9"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.720514 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9cxbm"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.733679 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9cxbm"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.746914 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.754289 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.766469 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 17:37:33 crc kubenswrapper[4809]: W1127 17:37:33.768102 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07137204_40d7_4004_a77d_6b704e1d1602.slice/crio-ceccb9d22686c2d8db5df424e123d28f91c3411f6407362cd75305af8361a96c WatchSource:0}: Error finding container ceccb9d22686c2d8db5df424e123d28f91c3411f6407362cd75305af8361a96c: Status 404 returned error can't find the container with id ceccb9d22686c2d8db5df424e123d28f91c3411f6407362cd75305af8361a96c Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.775646 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h8g6t"] Nov 27 17:37:33 crc kubenswrapper[4809]: I1127 17:37:33.946327 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 27 17:37:33 crc kubenswrapper[4809]: W1127 17:37:33.960577 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6ab5ce4_ce26_4eee_98fe_1bb380df9376.slice/crio-663ccdeea9f83c2caf015ef5c99d66f1525034e408e03adc341e8f59ee51e470 WatchSource:0}: Error finding container 663ccdeea9f83c2caf015ef5c99d66f1525034e408e03adc341e8f59ee51e470: Status 404 returned error can't find the container with id 663ccdeea9f83c2caf015ef5c99d66f1525034e408e03adc341e8f59ee51e470 Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.188644 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 27 17:37:34 crc kubenswrapper[4809]: W1127 17:37:34.211390 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcda8dbab_96fd_4d1b_a9cb_e056382df1a2.slice/crio-d3f60d73df1f2bb99be80b82d6f29a3180a5409d6b75d6b6616628242bdea544 WatchSource:0}: Error finding container d3f60d73df1f2bb99be80b82d6f29a3180a5409d6b75d6b6616628242bdea544: Status 404 returned error can't find the container with id d3f60d73df1f2bb99be80b82d6f29a3180a5409d6b75d6b6616628242bdea544 Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.585589 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782","Type":"ContainerStarted","Data":"d51363138159c3e3be6db7b3d4439c4f0eea33ae21bdb5dba4ecd68ed1de8c93"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.587619 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"07137204-40d7-4004-a77d-6b704e1d1602","Type":"ContainerStarted","Data":"ceccb9d22686c2d8db5df424e123d28f91c3411f6407362cd75305af8361a96c"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.589606 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01288244-f670-47a4-ac43-180191d7f331","Type":"ContainerStarted","Data":"8f46c1e32fcaae1fb5a8e6dc996aefc3138b25bd3a8696657ba6983c6225d736"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.591280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8145e23-a84e-405f-beb6-e27bb2aa1c59","Type":"ContainerStarted","Data":"7b776110c98052a76f6a2f9d39014d8cb0a8abc8d6c4405df4980798e25e8dd7"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.592534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bb517228-deb6-41fa-b6aa-354014cd34e1","Type":"ContainerStarted","Data":"ed5a66402078b658746f64f3a157b8d7de548389d651ebdb2c33921fd8ff9b8d"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.595041 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" event={"ID":"a67727fd-a1c7-4d34-897c-9522dd801a83","Type":"ContainerStarted","Data":"f5870dbafa14c920df512ab5e96b3ff2c97d5d76e35360db2ac62847756cb419"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.595241 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.600439 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a6ab5ce4-ce26-4eee-98fe-1bb380df9376","Type":"ContainerStarted","Data":"663ccdeea9f83c2caf015ef5c99d66f1525034e408e03adc341e8f59ee51e470"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.602152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jnvf9" event={"ID":"a501e4bf-f184-451f-9b21-a813f147413d","Type":"ContainerStarted","Data":"beef5440cd94d30fca7bb4330db3072229bfa7e088343b3f767497d82d368a3e"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.603264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cda8dbab-96fd-4d1b-a9cb-e056382df1a2","Type":"ContainerStarted","Data":"d3f60d73df1f2bb99be80b82d6f29a3180a5409d6b75d6b6616628242bdea544"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.604320 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t" event={"ID":"55d50ebb-8cab-42df-96b5-9598262337a4","Type":"ContainerStarted","Data":"28b2d4045cc58e53a77773ce8d17c8e04a1080fa51ab77642d30b4216ea76bc6"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.605790 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" event={"ID":"c9e56ebf-c6e1-4675-afcb-cbb600ac778e","Type":"ContainerStarted","Data":"a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a"} Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.606541 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.621130 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" podStartSLOduration=3.274074112 podStartE2EDuration="15.621115284s" podCreationTimestamp="2025-11-27 17:37:19 +0000 UTC" firstStartedPulling="2025-11-27 17:37:20.198992342 +0000 UTC m=+1675.471449694" lastFinishedPulling="2025-11-27 17:37:32.546033514 +0000 UTC m=+1687.818490866" observedRunningTime="2025-11-27 17:37:34.614609437 +0000 UTC m=+1689.887066789" watchObservedRunningTime="2025-11-27 17:37:34.621115284 +0000 UTC m=+1689.893572636" Nov 27 17:37:34 crc kubenswrapper[4809]: I1127 17:37:34.641137 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" podStartSLOduration=6.241181941 podStartE2EDuration="15.641053477s" podCreationTimestamp="2025-11-27 17:37:19 +0000 UTC" firstStartedPulling="2025-11-27 17:37:23.130355518 +0000 UTC m=+1678.402840411" lastFinishedPulling="2025-11-27 17:37:32.530254595 +0000 UTC m=+1687.802711947" observedRunningTime="2025-11-27 17:37:34.633954593 +0000 UTC m=+1689.906411955" watchObservedRunningTime="2025-11-27 17:37:34.641053477 +0000 UTC m=+1689.913510819" Nov 27 17:37:35 crc kubenswrapper[4809]: I1127 17:37:35.473211 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f" path="/var/lib/kubelet/pods/b6a7aea8-fcb6-4723-a025-6c3ce4c24a0f/volumes" Nov 27 17:37:35 crc kubenswrapper[4809]: I1127 17:37:35.473808 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d336542d-4cea-4b20-ae0b-c27ba218d364" path="/var/lib/kubelet/pods/d336542d-4cea-4b20-ae0b-c27ba218d364/volumes" Nov 27 17:37:39 crc kubenswrapper[4809]: I1127 17:37:39.675922 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:39 crc kubenswrapper[4809]: I1127 17:37:39.751176 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:37:39 crc kubenswrapper[4809]: I1127 17:37:39.813081 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-2qrtg"] Nov 27 17:37:39 crc kubenswrapper[4809]: I1127 17:37:39.813373 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" podUID="a67727fd-a1c7-4d34-897c-9522dd801a83" containerName="dnsmasq-dns" containerID="cri-o://f5870dbafa14c920df512ab5e96b3ff2c97d5d76e35360db2ac62847756cb419" gracePeriod=10 Nov 27 17:37:40 crc kubenswrapper[4809]: I1127 17:37:40.706615 4809 generic.go:334] "Generic (PLEG): container finished" podID="a67727fd-a1c7-4d34-897c-9522dd801a83" containerID="f5870dbafa14c920df512ab5e96b3ff2c97d5d76e35360db2ac62847756cb419" exitCode=0 Nov 27 17:37:40 crc kubenswrapper[4809]: I1127 17:37:40.706698 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" event={"ID":"a67727fd-a1c7-4d34-897c-9522dd801a83","Type":"ContainerDied","Data":"f5870dbafa14c920df512ab5e96b3ff2c97d5d76e35360db2ac62847756cb419"} Nov 27 17:37:42 crc kubenswrapper[4809]: I1127 17:37:42.823502 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:42 crc kubenswrapper[4809]: I1127 17:37:42.910468 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-config\") pod \"a67727fd-a1c7-4d34-897c-9522dd801a83\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " Nov 27 17:37:42 crc kubenswrapper[4809]: I1127 17:37:42.910708 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj24w\" (UniqueName: \"kubernetes.io/projected/a67727fd-a1c7-4d34-897c-9522dd801a83-kube-api-access-nj24w\") pod \"a67727fd-a1c7-4d34-897c-9522dd801a83\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " Nov 27 17:37:42 crc kubenswrapper[4809]: I1127 17:37:42.910769 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-dns-svc\") pod \"a67727fd-a1c7-4d34-897c-9522dd801a83\" (UID: \"a67727fd-a1c7-4d34-897c-9522dd801a83\") " Nov 27 17:37:42 crc kubenswrapper[4809]: I1127 17:37:42.914855 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a67727fd-a1c7-4d34-897c-9522dd801a83-kube-api-access-nj24w" (OuterVolumeSpecName: "kube-api-access-nj24w") pod "a67727fd-a1c7-4d34-897c-9522dd801a83" (UID: "a67727fd-a1c7-4d34-897c-9522dd801a83"). InnerVolumeSpecName "kube-api-access-nj24w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:37:42 crc kubenswrapper[4809]: I1127 17:37:42.958561 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-config" (OuterVolumeSpecName: "config") pod "a67727fd-a1c7-4d34-897c-9522dd801a83" (UID: "a67727fd-a1c7-4d34-897c-9522dd801a83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:37:42 crc kubenswrapper[4809]: I1127 17:37:42.958955 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a67727fd-a1c7-4d34-897c-9522dd801a83" (UID: "a67727fd-a1c7-4d34-897c-9522dd801a83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:37:43 crc kubenswrapper[4809]: I1127 17:37:43.012437 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj24w\" (UniqueName: \"kubernetes.io/projected/a67727fd-a1c7-4d34-897c-9522dd801a83-kube-api-access-nj24w\") on node \"crc\" DevicePath \"\"" Nov 27 17:37:43 crc kubenswrapper[4809]: I1127 17:37:43.012471 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:37:43 crc kubenswrapper[4809]: I1127 17:37:43.012480 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67727fd-a1c7-4d34-897c-9522dd801a83-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:37:43 crc kubenswrapper[4809]: I1127 17:37:43.730239 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" event={"ID":"a67727fd-a1c7-4d34-897c-9522dd801a83","Type":"ContainerDied","Data":"bbba2158a914c552a7466307b357dec59073058db0fa2c078ef0a4a88eb9fae0"} Nov 27 17:37:43 crc kubenswrapper[4809]: I1127 17:37:43.730298 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-2qrtg" Nov 27 17:37:43 crc kubenswrapper[4809]: I1127 17:37:43.730661 4809 scope.go:117] "RemoveContainer" containerID="f5870dbafa14c920df512ab5e96b3ff2c97d5d76e35360db2ac62847756cb419" Nov 27 17:37:43 crc kubenswrapper[4809]: I1127 17:37:43.748592 4809 scope.go:117] "RemoveContainer" containerID="af1456fc9d80c1fd1f5bae61d78fb8569bb7eac4e2d8facc31ca6c8e01789399" Nov 27 17:37:43 crc kubenswrapper[4809]: I1127 17:37:43.749478 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-2qrtg"] Nov 27 17:37:43 crc kubenswrapper[4809]: I1127 17:37:43.758090 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-2qrtg"] Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.739720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"07137204-40d7-4004-a77d-6b704e1d1602","Type":"ContainerStarted","Data":"9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4"} Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.740128 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.742864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t" event={"ID":"55d50ebb-8cab-42df-96b5-9598262337a4","Type":"ContainerStarted","Data":"f07fd1769bc53f36b85b0a8e3ea12d348ba1bcfbace1a95f738bba247515bacc"} Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.742969 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-h8g6t" Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.744312 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f82a4054-fea3-4071-b65d-529d900c0f04","Type":"ContainerStarted","Data":"b7c58b2a1b0e7bab27fe1c60ccebe522a84992f6a038ba36363581a6e7935af4"} Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.746363 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782","Type":"ContainerStarted","Data":"d7e3da1bf3eff3102a3d4d9aef547562fd7a620e2a431763497197193bcdaef3"} Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.748472 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bb517228-deb6-41fa-b6aa-354014cd34e1","Type":"ContainerStarted","Data":"f640d8443ace72018d3db68b4baecaa55f6d38dc3f31ecc010fc67b4fa8cef2f"} Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.748636 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.750064 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a6ab5ce4-ce26-4eee-98fe-1bb380df9376","Type":"ContainerStarted","Data":"01ae3cb02b20895b83420bd50b9facd8053e11f1faac06f701fa16712e03c985"} Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.751620 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jnvf9" event={"ID":"a501e4bf-f184-451f-9b21-a813f147413d","Type":"ContainerStarted","Data":"ac601a2de40a92bc8798230b57fb293e17b704c72b13ec3d60d749fbc3ca6b74"} Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.753439 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cda8dbab-96fd-4d1b-a9cb-e056382df1a2","Type":"ContainerStarted","Data":"20017f510fdd3ce62c3e8fe98a07bfd64c896aded0305b2572055f03e25be736"} Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.786346 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.218985467 podStartE2EDuration="19.786316281s" podCreationTimestamp="2025-11-27 17:37:25 +0000 UTC" firstStartedPulling="2025-11-27 17:37:33.776652049 +0000 UTC m=+1689.049109401" lastFinishedPulling="2025-11-27 17:37:43.343982863 +0000 UTC m=+1698.616440215" observedRunningTime="2025-11-27 17:37:44.756643594 +0000 UTC m=+1700.029100946" watchObservedRunningTime="2025-11-27 17:37:44.786316281 +0000 UTC m=+1700.058773633" Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.888080 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.936722013 podStartE2EDuration="21.88804956s" podCreationTimestamp="2025-11-27 17:37:23 +0000 UTC" firstStartedPulling="2025-11-27 17:37:33.765103635 +0000 UTC m=+1689.037560987" lastFinishedPulling="2025-11-27 17:37:42.716431182 +0000 UTC m=+1697.988888534" observedRunningTime="2025-11-27 17:37:44.873465493 +0000 UTC m=+1700.145922835" watchObservedRunningTime="2025-11-27 17:37:44.88804956 +0000 UTC m=+1700.160506912" Nov 27 17:37:44 crc kubenswrapper[4809]: I1127 17:37:44.901511 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-h8g6t" podStartSLOduration=6.841195072 podStartE2EDuration="15.901492795s" podCreationTimestamp="2025-11-27 17:37:29 +0000 UTC" firstStartedPulling="2025-11-27 17:37:33.788715458 +0000 UTC m=+1689.061172810" lastFinishedPulling="2025-11-27 17:37:42.849013181 +0000 UTC m=+1698.121470533" observedRunningTime="2025-11-27 17:37:44.891430382 +0000 UTC m=+1700.163887734" watchObservedRunningTime="2025-11-27 17:37:44.901492795 +0000 UTC m=+1700.173950147" Nov 27 17:37:45 crc kubenswrapper[4809]: I1127 17:37:45.463967 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:37:45 crc kubenswrapper[4809]: E1127 17:37:45.464523 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:37:45 crc kubenswrapper[4809]: I1127 17:37:45.472501 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a67727fd-a1c7-4d34-897c-9522dd801a83" path="/var/lib/kubelet/pods/a67727fd-a1c7-4d34-897c-9522dd801a83/volumes" Nov 27 17:37:45 crc kubenswrapper[4809]: I1127 17:37:45.763912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01288244-f670-47a4-ac43-180191d7f331","Type":"ContainerStarted","Data":"dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b"} Nov 27 17:37:45 crc kubenswrapper[4809]: I1127 17:37:45.767518 4809 generic.go:334] "Generic (PLEG): container finished" podID="a501e4bf-f184-451f-9b21-a813f147413d" containerID="ac601a2de40a92bc8798230b57fb293e17b704c72b13ec3d60d749fbc3ca6b74" exitCode=0 Nov 27 17:37:45 crc kubenswrapper[4809]: I1127 17:37:45.767574 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jnvf9" event={"ID":"a501e4bf-f184-451f-9b21-a813f147413d","Type":"ContainerDied","Data":"ac601a2de40a92bc8798230b57fb293e17b704c72b13ec3d60d749fbc3ca6b74"} Nov 27 17:37:45 crc kubenswrapper[4809]: I1127 17:37:45.773509 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8145e23-a84e-405f-beb6-e27bb2aa1c59","Type":"ContainerStarted","Data":"e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5"} Nov 27 17:37:49 crc kubenswrapper[4809]: I1127 17:37:49.193883 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 27 17:37:51 crc kubenswrapper[4809]: I1127 17:37:51.814699 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jnvf9" event={"ID":"a501e4bf-f184-451f-9b21-a813f147413d","Type":"ContainerStarted","Data":"09dd46d478d7f6715cb76567d6dec181255375f574abbc2be5ff0eff59098e71"} Nov 27 17:37:55 crc kubenswrapper[4809]: I1127 17:37:55.848795 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jnvf9" event={"ID":"a501e4bf-f184-451f-9b21-a813f147413d","Type":"ContainerStarted","Data":"10f7e5e2823e7edad295c7e11c76fd9e8c12cac3c1a21503d3850d80a676bf79"} Nov 27 17:37:55 crc kubenswrapper[4809]: I1127 17:37:55.849436 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:55 crc kubenswrapper[4809]: I1127 17:37:55.852276 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cda8dbab-96fd-4d1b-a9cb-e056382df1a2","Type":"ContainerStarted","Data":"4671c8e9cd7632bb32d34af64d5c4f81bfb24abc4961d5b22ddf63e9e0b76752"} Nov 27 17:37:55 crc kubenswrapper[4809]: I1127 17:37:55.855540 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a6ab5ce4-ce26-4eee-98fe-1bb380df9376","Type":"ContainerStarted","Data":"f9e624053b0a0a69a0a2ba02c2414fb8e9e03064d2b879d384c5791364dae735"} Nov 27 17:37:55 crc kubenswrapper[4809]: I1127 17:37:55.873703 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-jnvf9" podStartSLOduration=17.866243928 podStartE2EDuration="26.873683322s" podCreationTimestamp="2025-11-27 17:37:29 +0000 UTC" firstStartedPulling="2025-11-27 17:37:33.704943188 +0000 UTC m=+1688.977400550" lastFinishedPulling="2025-11-27 17:37:42.712382592 +0000 UTC m=+1697.984839944" observedRunningTime="2025-11-27 17:37:55.868291765 +0000 UTC m=+1711.140749137" watchObservedRunningTime="2025-11-27 17:37:55.873683322 +0000 UTC m=+1711.146140684" Nov 27 17:37:55 crc kubenswrapper[4809]: I1127 17:37:55.902951 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.020460205 podStartE2EDuration="24.902929138s" podCreationTimestamp="2025-11-27 17:37:31 +0000 UTC" firstStartedPulling="2025-11-27 17:37:33.965733266 +0000 UTC m=+1689.238190618" lastFinishedPulling="2025-11-27 17:37:54.848202199 +0000 UTC m=+1710.120659551" observedRunningTime="2025-11-27 17:37:55.889689337 +0000 UTC m=+1711.162146709" watchObservedRunningTime="2025-11-27 17:37:55.902929138 +0000 UTC m=+1711.175386490" Nov 27 17:37:55 crc kubenswrapper[4809]: I1127 17:37:55.920049 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.279460494 podStartE2EDuration="24.920024763s" podCreationTimestamp="2025-11-27 17:37:31 +0000 UTC" firstStartedPulling="2025-11-27 17:37:34.2149521 +0000 UTC m=+1689.487409452" lastFinishedPulling="2025-11-27 17:37:54.855516369 +0000 UTC m=+1710.127973721" observedRunningTime="2025-11-27 17:37:55.9118395 +0000 UTC m=+1711.184296852" watchObservedRunningTime="2025-11-27 17:37:55.920024763 +0000 UTC m=+1711.192482125" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.119789 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-zx94t"] Nov 27 17:37:56 crc kubenswrapper[4809]: E1127 17:37:56.120250 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a67727fd-a1c7-4d34-897c-9522dd801a83" containerName="init" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.120266 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a67727fd-a1c7-4d34-897c-9522dd801a83" containerName="init" Nov 27 17:37:56 crc kubenswrapper[4809]: E1127 17:37:56.120290 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a67727fd-a1c7-4d34-897c-9522dd801a83" containerName="dnsmasq-dns" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.120298 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a67727fd-a1c7-4d34-897c-9522dd801a83" containerName="dnsmasq-dns" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.120507 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a67727fd-a1c7-4d34-897c-9522dd801a83" containerName="dnsmasq-dns" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.121784 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.131609 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-zx94t"] Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.200082 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.250941 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-config\") pod \"dnsmasq-dns-7cb5889db5-zx94t\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.251001 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-zx94t\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.251087 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqd75\" (UniqueName: \"kubernetes.io/projected/c08a7860-cab4-4a97-9bd8-05dc5affed44-kube-api-access-lqd75\") pod \"dnsmasq-dns-7cb5889db5-zx94t\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.353947 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-config\") pod \"dnsmasq-dns-7cb5889db5-zx94t\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.354017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-zx94t\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.354118 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqd75\" (UniqueName: \"kubernetes.io/projected/c08a7860-cab4-4a97-9bd8-05dc5affed44-kube-api-access-lqd75\") pod \"dnsmasq-dns-7cb5889db5-zx94t\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.355160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-config\") pod \"dnsmasq-dns-7cb5889db5-zx94t\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.355350 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-zx94t\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.403872 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqd75\" (UniqueName: \"kubernetes.io/projected/c08a7860-cab4-4a97-9bd8-05dc5affed44-kube-api-access-lqd75\") pod \"dnsmasq-dns-7cb5889db5-zx94t\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.443167 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.457348 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:37:56 crc kubenswrapper[4809]: E1127 17:37:56.457683 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.865542 4809 generic.go:334] "Generic (PLEG): container finished" podID="f82a4054-fea3-4071-b65d-529d900c0f04" containerID="b7c58b2a1b0e7bab27fe1c60ccebe522a84992f6a038ba36363581a6e7935af4" exitCode=0 Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.865697 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f82a4054-fea3-4071-b65d-529d900c0f04","Type":"ContainerDied","Data":"b7c58b2a1b0e7bab27fe1c60ccebe522a84992f6a038ba36363581a6e7935af4"} Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.866130 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:37:56 crc kubenswrapper[4809]: W1127 17:37:56.896075 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc08a7860_cab4_4a97_9bd8_05dc5affed44.slice/crio-59b46fca2217c511e08a5a084553ee46d892061376a93de764c5f1313ca70fe8 WatchSource:0}: Error finding container 59b46fca2217c511e08a5a084553ee46d892061376a93de764c5f1313ca70fe8: Status 404 returned error can't find the container with id 59b46fca2217c511e08a5a084553ee46d892061376a93de764c5f1313ca70fe8 Nov 27 17:37:56 crc kubenswrapper[4809]: I1127 17:37:56.899497 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-zx94t"] Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.101340 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.139842 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.285627 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.290941 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.293334 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.293475 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.293556 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.293883 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-kw72w" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.310254 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.368191 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.372253 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.372377 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sv54\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-kube-api-access-4sv54\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.372399 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1b359048-c31b-4d16-922a-77b4b9afe87d-lock\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.372446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1b359048-c31b-4d16-922a-77b4b9afe87d-cache\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.372603 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.407563 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.474282 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.474406 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sv54\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-kube-api-access-4sv54\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.474437 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1b359048-c31b-4d16-922a-77b4b9afe87d-lock\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: E1127 17:37:57.474466 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 17:37:57 crc kubenswrapper[4809]: E1127 17:37:57.474495 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 17:37:57 crc kubenswrapper[4809]: E1127 17:37:57.474542 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift podName:1b359048-c31b-4d16-922a-77b4b9afe87d nodeName:}" failed. No retries permitted until 2025-11-27 17:37:57.974524534 +0000 UTC m=+1713.246981876 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift") pod "swift-storage-0" (UID: "1b359048-c31b-4d16-922a-77b4b9afe87d") : configmap "swift-ring-files" not found Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.474474 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1b359048-c31b-4d16-922a-77b4b9afe87d-cache\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.474664 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.475053 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.475075 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1b359048-c31b-4d16-922a-77b4b9afe87d-cache\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.475075 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1b359048-c31b-4d16-922a-77b4b9afe87d-lock\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.498144 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sv54\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-kube-api-access-4sv54\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.508099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.843947 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-hgrfv"] Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.845399 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.850252 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.850304 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.852368 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.866968 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hgrfv"] Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.882198 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f82a4054-fea3-4071-b65d-529d900c0f04","Type":"ContainerStarted","Data":"5f84158c6ff3aa7e1d384d834b9835d50e1c17854219f4da403f92787081e3cc"} Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.882542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-combined-ca-bundle\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.882602 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98c79518-5ef3-4382-a156-6de22fc92082-etc-swift\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.882638 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-scripts\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.882791 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-swiftconf\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.883486 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-dispersionconf\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.883554 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-ring-data-devices\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.883597 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfx9c\" (UniqueName: \"kubernetes.io/projected/98c79518-5ef3-4382-a156-6de22fc92082-kube-api-access-nfx9c\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.885317 4809 generic.go:334] "Generic (PLEG): container finished" podID="b3cf12fa-2068-4bcd-ae42-d7c6dfea6782" containerID="d7e3da1bf3eff3102a3d4d9aef547562fd7a620e2a431763497197193bcdaef3" exitCode=0 Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.885469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782","Type":"ContainerDied","Data":"d7e3da1bf3eff3102a3d4d9aef547562fd7a620e2a431763497197193bcdaef3"} Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.889368 4809 generic.go:334] "Generic (PLEG): container finished" podID="c08a7860-cab4-4a97-9bd8-05dc5affed44" containerID="5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03" exitCode=0 Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.890690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" event={"ID":"c08a7860-cab4-4a97-9bd8-05dc5affed44","Type":"ContainerDied","Data":"5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03"} Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.890756 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" event={"ID":"c08a7860-cab4-4a97-9bd8-05dc5affed44","Type":"ContainerStarted","Data":"59b46fca2217c511e08a5a084553ee46d892061376a93de764c5f1313ca70fe8"} Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.892047 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.892077 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.904834 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=26.653126023 podStartE2EDuration="35.904813645s" podCreationTimestamp="2025-11-27 17:37:22 +0000 UTC" firstStartedPulling="2025-11-27 17:37:33.464343369 +0000 UTC m=+1688.736800721" lastFinishedPulling="2025-11-27 17:37:42.716030991 +0000 UTC m=+1697.988488343" observedRunningTime="2025-11-27 17:37:57.90020391 +0000 UTC m=+1713.172661262" watchObservedRunningTime="2025-11-27 17:37:57.904813645 +0000 UTC m=+1713.177270997" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.941477 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.942369 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.985403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-combined-ca-bundle\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.985607 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98c79518-5ef3-4382-a156-6de22fc92082-etc-swift\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.985722 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-scripts\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.987675 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-swiftconf\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.988524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98c79518-5ef3-4382-a156-6de22fc92082-etc-swift\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.990235 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-scripts\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: E1127 17:37:57.990285 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 17:37:57 crc kubenswrapper[4809]: E1127 17:37:57.990478 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 17:37:57 crc kubenswrapper[4809]: E1127 17:37:57.990600 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift podName:1b359048-c31b-4d16-922a-77b4b9afe87d nodeName:}" failed. No retries permitted until 2025-11-27 17:37:58.99058086 +0000 UTC m=+1714.263038402 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift") pod "swift-storage-0" (UID: "1b359048-c31b-4d16-922a-77b4b9afe87d") : configmap "swift-ring-files" not found Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.988505 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.991423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-dispersionconf\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.991776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-ring-data-devices\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.991969 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfx9c\" (UniqueName: \"kubernetes.io/projected/98c79518-5ef3-4382-a156-6de22fc92082-kube-api-access-nfx9c\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:57 crc kubenswrapper[4809]: I1127 17:37:57.994720 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-ring-data-devices\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:57.999564 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-swiftconf\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.000270 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-combined-ca-bundle\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.003876 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-dispersionconf\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.018858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfx9c\" (UniqueName: \"kubernetes.io/projected/98c79518-5ef3-4382-a156-6de22fc92082-kube-api-access-nfx9c\") pod \"swift-ring-rebalance-hgrfv\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.133316 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-zx94t"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.163598 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.181933 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-pqrw8"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.186130 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.188191 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.197040 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.197143 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpqkq\" (UniqueName: \"kubernetes.io/projected/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-kube-api-access-bpqkq\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.197166 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-config\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.197191 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.203357 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-9567r"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.215278 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.221728 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-pqrw8"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.222623 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.229630 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9567r"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.302712 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.302772 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-config\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.302830 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.302857 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d27vh\" (UniqueName: \"kubernetes.io/projected/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-kube-api-access-d27vh\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.302881 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-ovn-rundir\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.302911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpqkq\" (UniqueName: \"kubernetes.io/projected/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-kube-api-access-bpqkq\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.302929 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-config\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.302952 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.302977 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-combined-ca-bundle\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.303005 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-ovs-rundir\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.303816 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.304541 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-config\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.305068 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.332116 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpqkq\" (UniqueName: \"kubernetes.io/projected/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-kube-api-access-bpqkq\") pod \"dnsmasq-dns-74f6f696b9-pqrw8\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.383190 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-pqrw8"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.384009 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.404766 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-config\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.404863 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.404887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d27vh\" (UniqueName: \"kubernetes.io/projected/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-kube-api-access-d27vh\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.404916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-ovn-rundir\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.404972 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-combined-ca-bundle\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.405000 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-ovs-rundir\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.405384 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-ovs-rundir\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.405434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-ovn-rundir\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.409396 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-config\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.433458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-combined-ca-bundle\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.434572 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.456177 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d27vh\" (UniqueName: \"kubernetes.io/projected/7a4f76ce-dff9-4d9c-a75b-7282bc630a36-kube-api-access-d27vh\") pod \"ovn-controller-metrics-9567r\" (UID: \"7a4f76ce-dff9-4d9c-a75b-7282bc630a36\") " pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.474154 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.476357 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.482717 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-lrl4h" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.482966 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.483200 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.483491 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.549808 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-wvhh5"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.551660 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.564207 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9567r" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.569546 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.569685 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.611094 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/545a6dbd-4f59-435b-b951-0a8e58be0d2a-scripts\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.611177 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/545a6dbd-4f59-435b-b951-0a8e58be0d2a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.611200 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl74f\" (UniqueName: \"kubernetes.io/projected/545a6dbd-4f59-435b-b951-0a8e58be0d2a-kube-api-access-fl74f\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.611279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545a6dbd-4f59-435b-b951-0a8e58be0d2a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.611305 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/545a6dbd-4f59-435b-b951-0a8e58be0d2a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.611363 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/545a6dbd-4f59-435b-b951-0a8e58be0d2a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.611384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/545a6dbd-4f59-435b-b951-0a8e58be0d2a-config\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.665471 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-wvhh5"] Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.713076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/545a6dbd-4f59-435b-b951-0a8e58be0d2a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.713532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/545a6dbd-4f59-435b-b951-0a8e58be0d2a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.713563 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/545a6dbd-4f59-435b-b951-0a8e58be0d2a-config\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.713632 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-config\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.713667 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt97w\" (UniqueName: \"kubernetes.io/projected/a4cefed0-40bf-451b-bc22-609d313f1d67-kube-api-access-qt97w\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.713700 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/545a6dbd-4f59-435b-b951-0a8e58be0d2a-scripts\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.713715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-dns-svc\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.723919 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/545a6dbd-4f59-435b-b951-0a8e58be0d2a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.723974 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl74f\" (UniqueName: \"kubernetes.io/projected/545a6dbd-4f59-435b-b951-0a8e58be0d2a-kube-api-access-fl74f\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.724078 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.724158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.724215 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545a6dbd-4f59-435b-b951-0a8e58be0d2a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.725629 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/545a6dbd-4f59-435b-b951-0a8e58be0d2a-config\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.726414 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/545a6dbd-4f59-435b-b951-0a8e58be0d2a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.727137 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/545a6dbd-4f59-435b-b951-0a8e58be0d2a-scripts\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.727370 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/545a6dbd-4f59-435b-b951-0a8e58be0d2a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.733280 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/545a6dbd-4f59-435b-b951-0a8e58be0d2a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.743029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545a6dbd-4f59-435b-b951-0a8e58be0d2a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.780504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl74f\" (UniqueName: \"kubernetes.io/projected/545a6dbd-4f59-435b-b951-0a8e58be0d2a-kube-api-access-fl74f\") pod \"ovn-northd-0\" (UID: \"545a6dbd-4f59-435b-b951-0a8e58be0d2a\") " pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.827035 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt97w\" (UniqueName: \"kubernetes.io/projected/a4cefed0-40bf-451b-bc22-609d313f1d67-kube-api-access-qt97w\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.827091 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-dns-svc\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.827152 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.827185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.827276 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-config\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.828643 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-config\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.829936 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-dns-svc\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.830618 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.833974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.856171 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt97w\" (UniqueName: \"kubernetes.io/projected/a4cefed0-40bf-451b-bc22-609d313f1d67-kube-api-access-qt97w\") pod \"dnsmasq-dns-698758b865-wvhh5\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.908732 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b3cf12fa-2068-4bcd-ae42-d7c6dfea6782","Type":"ContainerStarted","Data":"01aeedea4797775f4d726bd0c8145ef2cc280b7d0b3a803954d16f6a84896910"} Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.913937 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" event={"ID":"c08a7860-cab4-4a97-9bd8-05dc5affed44","Type":"ContainerStarted","Data":"9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84"} Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.927801 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=28.427733336 podStartE2EDuration="37.927785568s" podCreationTimestamp="2025-11-27 17:37:21 +0000 UTC" firstStartedPulling="2025-11-27 17:37:33.749530682 +0000 UTC m=+1689.021988034" lastFinishedPulling="2025-11-27 17:37:43.249582914 +0000 UTC m=+1698.522040266" observedRunningTime="2025-11-27 17:37:58.927300775 +0000 UTC m=+1714.199758127" watchObservedRunningTime="2025-11-27 17:37:58.927785568 +0000 UTC m=+1714.200242920" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.939116 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 27 17:37:58 crc kubenswrapper[4809]: I1127 17:37:58.964168 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" podStartSLOduration=2.9641503780000003 podStartE2EDuration="2.964150378s" podCreationTimestamp="2025-11-27 17:37:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:37:58.954858205 +0000 UTC m=+1714.227315577" watchObservedRunningTime="2025-11-27 17:37:58.964150378 +0000 UTC m=+1714.236607730" Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.014720 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.037933 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:37:59 crc kubenswrapper[4809]: E1127 17:37:59.051314 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 17:37:59 crc kubenswrapper[4809]: E1127 17:37:59.051335 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 17:37:59 crc kubenswrapper[4809]: E1127 17:37:59.051376 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift podName:1b359048-c31b-4d16-922a-77b4b9afe87d nodeName:}" failed. No retries permitted until 2025-11-27 17:38:01.051359691 +0000 UTC m=+1716.323817043 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift") pod "swift-storage-0" (UID: "1b359048-c31b-4d16-922a-77b4b9afe87d") : configmap "swift-ring-files" not found Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.187026 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hgrfv"] Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.202661 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-pqrw8"] Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.344401 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9567r"] Nov 27 17:37:59 crc kubenswrapper[4809]: W1127 17:37:59.358474 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a4f76ce_dff9_4d9c_a75b_7282bc630a36.slice/crio-6ec5ffc4e7233671639cc48cc7f556a452d25ca51d4e969f6475a3924b9943ae WatchSource:0}: Error finding container 6ec5ffc4e7233671639cc48cc7f556a452d25ca51d4e969f6475a3924b9943ae: Status 404 returned error can't find the container with id 6ec5ffc4e7233671639cc48cc7f556a452d25ca51d4e969f6475a3924b9943ae Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.529152 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-wvhh5"] Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.538561 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.921439 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"545a6dbd-4f59-435b-b951-0a8e58be0d2a","Type":"ContainerStarted","Data":"11ab1c62a2808310bc7907528b9fee6c35503fad6be77f6a285b24e218eb2e99"} Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.924306 4809 generic.go:334] "Generic (PLEG): container finished" podID="c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f" containerID="f5a44a8305ec1393554a65b8590f905bb4c040d0772a8f98bb7efc53df60951c" exitCode=0 Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.924771 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" event={"ID":"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f","Type":"ContainerDied","Data":"f5a44a8305ec1393554a65b8590f905bb4c040d0772a8f98bb7efc53df60951c"} Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.924819 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" event={"ID":"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f","Type":"ContainerStarted","Data":"62a8b0e53ca6a8ecd80f6a556ae0d03b1f46b65ddd6f6c9677d08851cb4e00bc"} Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.927529 4809 generic.go:334] "Generic (PLEG): container finished" podID="a4cefed0-40bf-451b-bc22-609d313f1d67" containerID="8bcd7f10799eb9e4a1bc14f47ab770911f873ea6a7a03f47e203b16cf4c55867" exitCode=0 Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.927614 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-wvhh5" event={"ID":"a4cefed0-40bf-451b-bc22-609d313f1d67","Type":"ContainerDied","Data":"8bcd7f10799eb9e4a1bc14f47ab770911f873ea6a7a03f47e203b16cf4c55867"} Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.927639 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-wvhh5" event={"ID":"a4cefed0-40bf-451b-bc22-609d313f1d67","Type":"ContainerStarted","Data":"818b5edf05f854e47b5feaf828f85cbddcd8aa65840cd1cb16f4dd3e2dfb0e16"} Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.932672 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hgrfv" event={"ID":"98c79518-5ef3-4382-a156-6de22fc92082","Type":"ContainerStarted","Data":"1a1c4f63d4595f9ee9c813340224a79d1ef38fea369dd7090a650a60aef1ad68"} Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.944479 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" podUID="c08a7860-cab4-4a97-9bd8-05dc5affed44" containerName="dnsmasq-dns" containerID="cri-o://9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84" gracePeriod=10 Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.945597 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9567r" event={"ID":"7a4f76ce-dff9-4d9c-a75b-7282bc630a36","Type":"ContainerStarted","Data":"b67bea4c1c1896bbf777de6e0ad49d88072480c98a428cdd0ecaedae69193125"} Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.945622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9567r" event={"ID":"7a4f76ce-dff9-4d9c-a75b-7282bc630a36","Type":"ContainerStarted","Data":"6ec5ffc4e7233671639cc48cc7f556a452d25ca51d4e969f6475a3924b9943ae"} Nov 27 17:37:59 crc kubenswrapper[4809]: I1127 17:37:59.946752 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.020996 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-9567r" podStartSLOduration=2.020970233 podStartE2EDuration="2.020970233s" podCreationTimestamp="2025-11-27 17:37:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:00.007664631 +0000 UTC m=+1715.280121983" watchObservedRunningTime="2025-11-27 17:38:00.020970233 +0000 UTC m=+1715.293427585" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.313553 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.379077 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-config\") pod \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.379670 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-dns-svc\") pod \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.379704 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpqkq\" (UniqueName: \"kubernetes.io/projected/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-kube-api-access-bpqkq\") pod \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.379753 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-ovsdbserver-nb\") pod \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\" (UID: \"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f\") " Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.384166 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-kube-api-access-bpqkq" (OuterVolumeSpecName: "kube-api-access-bpqkq") pod "c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f" (UID: "c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f"). InnerVolumeSpecName "kube-api-access-bpqkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.409329 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f" (UID: "c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.418306 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-config" (OuterVolumeSpecName: "config") pod "c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f" (UID: "c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.436103 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f" (UID: "c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.482360 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.482400 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpqkq\" (UniqueName: \"kubernetes.io/projected/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-kube-api-access-bpqkq\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.482416 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.482428 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.523373 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.583770 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-dns-svc\") pod \"c08a7860-cab4-4a97-9bd8-05dc5affed44\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.583879 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqd75\" (UniqueName: \"kubernetes.io/projected/c08a7860-cab4-4a97-9bd8-05dc5affed44-kube-api-access-lqd75\") pod \"c08a7860-cab4-4a97-9bd8-05dc5affed44\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.583926 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-config\") pod \"c08a7860-cab4-4a97-9bd8-05dc5affed44\" (UID: \"c08a7860-cab4-4a97-9bd8-05dc5affed44\") " Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.593015 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c08a7860-cab4-4a97-9bd8-05dc5affed44-kube-api-access-lqd75" (OuterVolumeSpecName: "kube-api-access-lqd75") pod "c08a7860-cab4-4a97-9bd8-05dc5affed44" (UID: "c08a7860-cab4-4a97-9bd8-05dc5affed44"). InnerVolumeSpecName "kube-api-access-lqd75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.661640 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c08a7860-cab4-4a97-9bd8-05dc5affed44" (UID: "c08a7860-cab4-4a97-9bd8-05dc5affed44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.664325 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-config" (OuterVolumeSpecName: "config") pod "c08a7860-cab4-4a97-9bd8-05dc5affed44" (UID: "c08a7860-cab4-4a97-9bd8-05dc5affed44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.686261 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.686297 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqd75\" (UniqueName: \"kubernetes.io/projected/c08a7860-cab4-4a97-9bd8-05dc5affed44-kube-api-access-lqd75\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.686312 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c08a7860-cab4-4a97-9bd8-05dc5affed44-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.961035 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" event={"ID":"c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f","Type":"ContainerDied","Data":"62a8b0e53ca6a8ecd80f6a556ae0d03b1f46b65ddd6f6c9677d08851cb4e00bc"} Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.961520 4809 scope.go:117] "RemoveContainer" containerID="f5a44a8305ec1393554a65b8590f905bb4c040d0772a8f98bb7efc53df60951c" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.961070 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-pqrw8" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.970857 4809 generic.go:334] "Generic (PLEG): container finished" podID="c08a7860-cab4-4a97-9bd8-05dc5affed44" containerID="9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84" exitCode=0 Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.970920 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" event={"ID":"c08a7860-cab4-4a97-9bd8-05dc5affed44","Type":"ContainerDied","Data":"9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84"} Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.970938 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" event={"ID":"c08a7860-cab4-4a97-9bd8-05dc5affed44","Type":"ContainerDied","Data":"59b46fca2217c511e08a5a084553ee46d892061376a93de764c5f1313ca70fe8"} Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.970976 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-zx94t" Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.974005 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-wvhh5" event={"ID":"a4cefed0-40bf-451b-bc22-609d313f1d67","Type":"ContainerStarted","Data":"86cd05b94e76ae5473fc96506b128cf3eb2af7f27861053ba3d0d7e6e2e362e1"} Nov 27 17:38:00 crc kubenswrapper[4809]: I1127 17:38:00.974229 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:38:01 crc kubenswrapper[4809]: I1127 17:38:01.007514 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-wvhh5" podStartSLOduration=3.007490504 podStartE2EDuration="3.007490504s" podCreationTimestamp="2025-11-27 17:37:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:00.999212479 +0000 UTC m=+1716.271669861" watchObservedRunningTime="2025-11-27 17:38:01.007490504 +0000 UTC m=+1716.279947866" Nov 27 17:38:01 crc kubenswrapper[4809]: I1127 17:38:01.030192 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-zx94t"] Nov 27 17:38:01 crc kubenswrapper[4809]: I1127 17:38:01.039259 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-zx94t"] Nov 27 17:38:01 crc kubenswrapper[4809]: I1127 17:38:01.057848 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-pqrw8"] Nov 27 17:38:01 crc kubenswrapper[4809]: I1127 17:38:01.065872 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-pqrw8"] Nov 27 17:38:01 crc kubenswrapper[4809]: I1127 17:38:01.105952 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:38:01 crc kubenswrapper[4809]: E1127 17:38:01.107999 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 17:38:01 crc kubenswrapper[4809]: E1127 17:38:01.108020 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 17:38:01 crc kubenswrapper[4809]: E1127 17:38:01.108056 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift podName:1b359048-c31b-4d16-922a-77b4b9afe87d nodeName:}" failed. No retries permitted until 2025-11-27 17:38:05.108040701 +0000 UTC m=+1720.380498053 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift") pod "swift-storage-0" (UID: "1b359048-c31b-4d16-922a-77b4b9afe87d") : configmap "swift-ring-files" not found Nov 27 17:38:01 crc kubenswrapper[4809]: I1127 17:38:01.468551 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c08a7860-cab4-4a97-9bd8-05dc5affed44" path="/var/lib/kubelet/pods/c08a7860-cab4-4a97-9bd8-05dc5affed44/volumes" Nov 27 17:38:01 crc kubenswrapper[4809]: I1127 17:38:01.469474 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f" path="/var/lib/kubelet/pods/c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f/volumes" Nov 27 17:38:02 crc kubenswrapper[4809]: I1127 17:38:02.530359 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 27 17:38:02 crc kubenswrapper[4809]: I1127 17:38:02.530440 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 27 17:38:03 crc kubenswrapper[4809]: I1127 17:38:03.765882 4809 scope.go:117] "RemoveContainer" containerID="9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84" Nov 27 17:38:03 crc kubenswrapper[4809]: I1127 17:38:03.841937 4809 scope.go:117] "RemoveContainer" containerID="5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03" Nov 27 17:38:04 crc kubenswrapper[4809]: I1127 17:38:04.027644 4809 scope.go:117] "RemoveContainer" containerID="9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84" Nov 27 17:38:04 crc kubenswrapper[4809]: E1127 17:38:04.028433 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84\": container with ID starting with 9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84 not found: ID does not exist" containerID="9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84" Nov 27 17:38:04 crc kubenswrapper[4809]: I1127 17:38:04.028471 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84"} err="failed to get container status \"9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84\": rpc error: code = NotFound desc = could not find container \"9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84\": container with ID starting with 9d28b12060eaba00a8db13f469a6dffd9dcd4a1dc0929d779aaa010a9d3fad84 not found: ID does not exist" Nov 27 17:38:04 crc kubenswrapper[4809]: I1127 17:38:04.028538 4809 scope.go:117] "RemoveContainer" containerID="5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03" Nov 27 17:38:04 crc kubenswrapper[4809]: E1127 17:38:04.033860 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03\": container with ID starting with 5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03 not found: ID does not exist" containerID="5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03" Nov 27 17:38:04 crc kubenswrapper[4809]: I1127 17:38:04.033937 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03"} err="failed to get container status \"5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03\": rpc error: code = NotFound desc = could not find container \"5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03\": container with ID starting with 5d66becdf3c80332e0017cb3303f6a7eb05154649e6fb3f336c6d6d3b6f98b03 not found: ID does not exist" Nov 27 17:38:04 crc kubenswrapper[4809]: I1127 17:38:04.316702 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 27 17:38:04 crc kubenswrapper[4809]: I1127 17:38:04.317092 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 27 17:38:04 crc kubenswrapper[4809]: I1127 17:38:04.392608 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 27 17:38:04 crc kubenswrapper[4809]: I1127 17:38:04.929924 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 27 17:38:05 crc kubenswrapper[4809]: I1127 17:38:05.008911 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 27 17:38:05 crc kubenswrapper[4809]: I1127 17:38:05.047805 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hgrfv" event={"ID":"98c79518-5ef3-4382-a156-6de22fc92082","Type":"ContainerStarted","Data":"3d9c1ebc424edeb26aabcb426cbf23ad2275f82fea72fc461210210a734feee6"} Nov 27 17:38:05 crc kubenswrapper[4809]: I1127 17:38:05.049902 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"545a6dbd-4f59-435b-b951-0a8e58be0d2a","Type":"ContainerStarted","Data":"71adb801a7caf27f83685257ae55ce0bb73a85965e5b0f0e876fe98980adec78"} Nov 27 17:38:05 crc kubenswrapper[4809]: I1127 17:38:05.050084 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"545a6dbd-4f59-435b-b951-0a8e58be0d2a","Type":"ContainerStarted","Data":"23d93d2386aaac6fce9f604e1cfe4aaa134388f56c70ae359c38f9f84eb5572c"} Nov 27 17:38:05 crc kubenswrapper[4809]: I1127 17:38:05.078912 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-hgrfv" podStartSLOduration=3.433496321 podStartE2EDuration="8.078890599s" podCreationTimestamp="2025-11-27 17:37:57 +0000 UTC" firstStartedPulling="2025-11-27 17:37:59.197660804 +0000 UTC m=+1714.470118156" lastFinishedPulling="2025-11-27 17:38:03.843055082 +0000 UTC m=+1719.115512434" observedRunningTime="2025-11-27 17:38:05.068111776 +0000 UTC m=+1720.340569128" watchObservedRunningTime="2025-11-27 17:38:05.078890599 +0000 UTC m=+1720.351347961" Nov 27 17:38:05 crc kubenswrapper[4809]: I1127 17:38:05.115317 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=5.771460313 podStartE2EDuration="7.11529541s" podCreationTimestamp="2025-11-27 17:37:58 +0000 UTC" firstStartedPulling="2025-11-27 17:37:59.545718357 +0000 UTC m=+1714.818175709" lastFinishedPulling="2025-11-27 17:38:00.889553454 +0000 UTC m=+1716.162010806" observedRunningTime="2025-11-27 17:38:05.111016564 +0000 UTC m=+1720.383473916" watchObservedRunningTime="2025-11-27 17:38:05.11529541 +0000 UTC m=+1720.387752762" Nov 27 17:38:05 crc kubenswrapper[4809]: I1127 17:38:05.145885 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 27 17:38:05 crc kubenswrapper[4809]: I1127 17:38:05.188554 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:38:05 crc kubenswrapper[4809]: E1127 17:38:05.189889 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 17:38:05 crc kubenswrapper[4809]: E1127 17:38:05.189909 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 17:38:05 crc kubenswrapper[4809]: E1127 17:38:05.189942 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift podName:1b359048-c31b-4d16-922a-77b4b9afe87d nodeName:}" failed. No retries permitted until 2025-11-27 17:38:13.189929032 +0000 UTC m=+1728.462386384 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift") pod "swift-storage-0" (UID: "1b359048-c31b-4d16-922a-77b4b9afe87d") : configmap "swift-ring-files" not found Nov 27 17:38:06 crc kubenswrapper[4809]: I1127 17:38:06.057459 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.016922 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.071794 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pmp8v"] Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.072019 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" podUID="c9e56ebf-c6e1-4675-afcb-cbb600ac778e" containerName="dnsmasq-dns" containerID="cri-o://a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a" gracePeriod=10 Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.436593 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-54b8-account-create-update-4t2zg"] Nov 27 17:38:09 crc kubenswrapper[4809]: E1127 17:38:09.437104 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f" containerName="init" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.437126 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f" containerName="init" Nov 27 17:38:09 crc kubenswrapper[4809]: E1127 17:38:09.437154 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08a7860-cab4-4a97-9bd8-05dc5affed44" containerName="init" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.437165 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08a7860-cab4-4a97-9bd8-05dc5affed44" containerName="init" Nov 27 17:38:09 crc kubenswrapper[4809]: E1127 17:38:09.437177 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08a7860-cab4-4a97-9bd8-05dc5affed44" containerName="dnsmasq-dns" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.437185 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08a7860-cab4-4a97-9bd8-05dc5affed44" containerName="dnsmasq-dns" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.437613 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0cfcc6b-1d36-4efd-bfe8-f74ebeed6c7f" containerName="init" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.437639 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c08a7860-cab4-4a97-9bd8-05dc5affed44" containerName="dnsmasq-dns" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.438384 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.440711 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.448206 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ql6vt"] Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.449637 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.468833 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2slq\" (UniqueName: \"kubernetes.io/projected/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-kube-api-access-r2slq\") pod \"glance-db-create-ql6vt\" (UID: \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\") " pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.468891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvl87\" (UniqueName: \"kubernetes.io/projected/91f8d060-b713-4586-b388-c68c51647297-kube-api-access-zvl87\") pod \"glance-54b8-account-create-update-4t2zg\" (UID: \"91f8d060-b713-4586-b388-c68c51647297\") " pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.468915 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-operator-scripts\") pod \"glance-db-create-ql6vt\" (UID: \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\") " pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.468969 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91f8d060-b713-4586-b388-c68c51647297-operator-scripts\") pod \"glance-54b8-account-create-update-4t2zg\" (UID: \"91f8d060-b713-4586-b388-c68c51647297\") " pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.473633 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-54b8-account-create-update-4t2zg"] Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.481507 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ql6vt"] Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.570790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2slq\" (UniqueName: \"kubernetes.io/projected/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-kube-api-access-r2slq\") pod \"glance-db-create-ql6vt\" (UID: \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\") " pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.570855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvl87\" (UniqueName: \"kubernetes.io/projected/91f8d060-b713-4586-b388-c68c51647297-kube-api-access-zvl87\") pod \"glance-54b8-account-create-update-4t2zg\" (UID: \"91f8d060-b713-4586-b388-c68c51647297\") " pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.570883 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-operator-scripts\") pod \"glance-db-create-ql6vt\" (UID: \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\") " pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.570943 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91f8d060-b713-4586-b388-c68c51647297-operator-scripts\") pod \"glance-54b8-account-create-update-4t2zg\" (UID: \"91f8d060-b713-4586-b388-c68c51647297\") " pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.571888 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91f8d060-b713-4586-b388-c68c51647297-operator-scripts\") pod \"glance-54b8-account-create-update-4t2zg\" (UID: \"91f8d060-b713-4586-b388-c68c51647297\") " pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.571973 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-operator-scripts\") pod \"glance-db-create-ql6vt\" (UID: \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\") " pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.583507 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.592975 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2slq\" (UniqueName: \"kubernetes.io/projected/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-kube-api-access-r2slq\") pod \"glance-db-create-ql6vt\" (UID: \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\") " pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.595999 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvl87\" (UniqueName: \"kubernetes.io/projected/91f8d060-b713-4586-b388-c68c51647297-kube-api-access-zvl87\") pod \"glance-54b8-account-create-update-4t2zg\" (UID: \"91f8d060-b713-4586-b388-c68c51647297\") " pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.671998 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-config\") pod \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.672128 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf5zg\" (UniqueName: \"kubernetes.io/projected/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-kube-api-access-sf5zg\") pod \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.672169 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-dns-svc\") pod \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\" (UID: \"c9e56ebf-c6e1-4675-afcb-cbb600ac778e\") " Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.684999 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-kube-api-access-sf5zg" (OuterVolumeSpecName: "kube-api-access-sf5zg") pod "c9e56ebf-c6e1-4675-afcb-cbb600ac778e" (UID: "c9e56ebf-c6e1-4675-afcb-cbb600ac778e"). InnerVolumeSpecName "kube-api-access-sf5zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.711420 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-config" (OuterVolumeSpecName: "config") pod "c9e56ebf-c6e1-4675-afcb-cbb600ac778e" (UID: "c9e56ebf-c6e1-4675-afcb-cbb600ac778e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.714302 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9e56ebf-c6e1-4675-afcb-cbb600ac778e" (UID: "c9e56ebf-c6e1-4675-afcb-cbb600ac778e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.774407 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf5zg\" (UniqueName: \"kubernetes.io/projected/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-kube-api-access-sf5zg\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.774437 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.774446 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e56ebf-c6e1-4675-afcb-cbb600ac778e-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.774551 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:09 crc kubenswrapper[4809]: I1127 17:38:09.785147 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.097637 4809 generic.go:334] "Generic (PLEG): container finished" podID="c9e56ebf-c6e1-4675-afcb-cbb600ac778e" containerID="a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a" exitCode=0 Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.097692 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" event={"ID":"c9e56ebf-c6e1-4675-afcb-cbb600ac778e","Type":"ContainerDied","Data":"a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a"} Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.097728 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" event={"ID":"c9e56ebf-c6e1-4675-afcb-cbb600ac778e","Type":"ContainerDied","Data":"7823488e5020c66057c05bc12141c4300559609ef470f7313a1fc0ed5b07e550"} Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.097773 4809 scope.go:117] "RemoveContainer" containerID="a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a" Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.097801 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-pmp8v" Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.133182 4809 scope.go:117] "RemoveContainer" containerID="503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72" Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.136699 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pmp8v"] Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.161767 4809 scope.go:117] "RemoveContainer" containerID="a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a" Nov 27 17:38:10 crc kubenswrapper[4809]: E1127 17:38:10.162177 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a\": container with ID starting with a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a not found: ID does not exist" containerID="a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a" Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.162207 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a"} err="failed to get container status \"a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a\": rpc error: code = NotFound desc = could not find container \"a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a\": container with ID starting with a4a9936e51724e50d0aa3db01d5d77e07ee1744774cce1d30e9a28d871fb318a not found: ID does not exist" Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.162227 4809 scope.go:117] "RemoveContainer" containerID="503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72" Nov 27 17:38:10 crc kubenswrapper[4809]: E1127 17:38:10.162639 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72\": container with ID starting with 503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72 not found: ID does not exist" containerID="503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72" Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.162704 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72"} err="failed to get container status \"503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72\": rpc error: code = NotFound desc = could not find container \"503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72\": container with ID starting with 503a0285e4911dddcb6914967a0e699e5fa4527f55c31478eaf478445fa7fc72 not found: ID does not exist" Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.169622 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-pmp8v"] Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.393344 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-54b8-account-create-update-4t2zg"] Nov 27 17:38:10 crc kubenswrapper[4809]: W1127 17:38:10.460053 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16f2aca6_31a1_41d6_9fbb_1a3fd8ff3f8f.slice/crio-dc46d8df8b54a853fd313727feed3e4d80d0a9000f281a70c350424c30b8ea6c WatchSource:0}: Error finding container dc46d8df8b54a853fd313727feed3e4d80d0a9000f281a70c350424c30b8ea6c: Status 404 returned error can't find the container with id dc46d8df8b54a853fd313727feed3e4d80d0a9000f281a70c350424c30b8ea6c Nov 27 17:38:10 crc kubenswrapper[4809]: I1127 17:38:10.461705 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ql6vt"] Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.107412 4809 generic.go:334] "Generic (PLEG): container finished" podID="91f8d060-b713-4586-b388-c68c51647297" containerID="478b94db3d1c86917b78ae47a4fa5c9dbfcdd36b291df6e8312989a838f2370c" exitCode=0 Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.107509 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-54b8-account-create-update-4t2zg" event={"ID":"91f8d060-b713-4586-b388-c68c51647297","Type":"ContainerDied","Data":"478b94db3d1c86917b78ae47a4fa5c9dbfcdd36b291df6e8312989a838f2370c"} Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.107865 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-54b8-account-create-update-4t2zg" event={"ID":"91f8d060-b713-4586-b388-c68c51647297","Type":"ContainerStarted","Data":"7f3e1be2dc9f2596a5e0005a1474caf6f4def899e24974ae78904347b3a31997"} Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.109713 4809 generic.go:334] "Generic (PLEG): container finished" podID="98c79518-5ef3-4382-a156-6de22fc92082" containerID="3d9c1ebc424edeb26aabcb426cbf23ad2275f82fea72fc461210210a734feee6" exitCode=0 Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.109805 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hgrfv" event={"ID":"98c79518-5ef3-4382-a156-6de22fc92082","Type":"ContainerDied","Data":"3d9c1ebc424edeb26aabcb426cbf23ad2275f82fea72fc461210210a734feee6"} Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.112426 4809 generic.go:334] "Generic (PLEG): container finished" podID="16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f" containerID="8c59fd8c4df8f089ddbfa38ce89e026badace6a1d40a425fd77c62ab417cab93" exitCode=0 Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.112460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ql6vt" event={"ID":"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f","Type":"ContainerDied","Data":"8c59fd8c4df8f089ddbfa38ce89e026badace6a1d40a425fd77c62ab417cab93"} Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.112483 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ql6vt" event={"ID":"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f","Type":"ContainerStarted","Data":"dc46d8df8b54a853fd313727feed3e4d80d0a9000f281a70c350424c30b8ea6c"} Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.458186 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:38:11 crc kubenswrapper[4809]: E1127 17:38:11.458470 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:38:11 crc kubenswrapper[4809]: I1127 17:38:11.469123 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e56ebf-c6e1-4675-afcb-cbb600ac778e" path="/var/lib/kubelet/pods/c9e56ebf-c6e1-4675-afcb-cbb600ac778e/volumes" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.571145 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.579970 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.594119 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729550 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2slq\" (UniqueName: \"kubernetes.io/projected/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-kube-api-access-r2slq\") pod \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\" (UID: \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729598 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-ring-data-devices\") pod \"98c79518-5ef3-4382-a156-6de22fc92082\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729616 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91f8d060-b713-4586-b388-c68c51647297-operator-scripts\") pod \"91f8d060-b713-4586-b388-c68c51647297\" (UID: \"91f8d060-b713-4586-b388-c68c51647297\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729659 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-combined-ca-bundle\") pod \"98c79518-5ef3-4382-a156-6de22fc92082\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729699 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfx9c\" (UniqueName: \"kubernetes.io/projected/98c79518-5ef3-4382-a156-6de22fc92082-kube-api-access-nfx9c\") pod \"98c79518-5ef3-4382-a156-6de22fc92082\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729718 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvl87\" (UniqueName: \"kubernetes.io/projected/91f8d060-b713-4586-b388-c68c51647297-kube-api-access-zvl87\") pod \"91f8d060-b713-4586-b388-c68c51647297\" (UID: \"91f8d060-b713-4586-b388-c68c51647297\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729755 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98c79518-5ef3-4382-a156-6de22fc92082-etc-swift\") pod \"98c79518-5ef3-4382-a156-6de22fc92082\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729784 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-scripts\") pod \"98c79518-5ef3-4382-a156-6de22fc92082\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729810 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-operator-scripts\") pod \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\" (UID: \"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.729847 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-dispersionconf\") pod \"98c79518-5ef3-4382-a156-6de22fc92082\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.730432 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "98c79518-5ef3-4382-a156-6de22fc92082" (UID: "98c79518-5ef3-4382-a156-6de22fc92082"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.730463 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-swiftconf\") pod \"98c79518-5ef3-4382-a156-6de22fc92082\" (UID: \"98c79518-5ef3-4382-a156-6de22fc92082\") " Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.730463 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91f8d060-b713-4586-b388-c68c51647297-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "91f8d060-b713-4586-b388-c68c51647297" (UID: "91f8d060-b713-4586-b388-c68c51647297"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.730901 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f" (UID: "16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.731084 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c79518-5ef3-4382-a156-6de22fc92082-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "98c79518-5ef3-4382-a156-6de22fc92082" (UID: "98c79518-5ef3-4382-a156-6de22fc92082"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.731188 4809 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.731252 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91f8d060-b713-4586-b388-c68c51647297-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.731265 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.735643 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-kube-api-access-r2slq" (OuterVolumeSpecName: "kube-api-access-r2slq") pod "16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f" (UID: "16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f"). InnerVolumeSpecName "kube-api-access-r2slq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.736079 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98c79518-5ef3-4382-a156-6de22fc92082-kube-api-access-nfx9c" (OuterVolumeSpecName: "kube-api-access-nfx9c") pod "98c79518-5ef3-4382-a156-6de22fc92082" (UID: "98c79518-5ef3-4382-a156-6de22fc92082"). InnerVolumeSpecName "kube-api-access-nfx9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.736327 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f8d060-b713-4586-b388-c68c51647297-kube-api-access-zvl87" (OuterVolumeSpecName: "kube-api-access-zvl87") pod "91f8d060-b713-4586-b388-c68c51647297" (UID: "91f8d060-b713-4586-b388-c68c51647297"). InnerVolumeSpecName "kube-api-access-zvl87". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.739989 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "98c79518-5ef3-4382-a156-6de22fc92082" (UID: "98c79518-5ef3-4382-a156-6de22fc92082"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.758557 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-scripts" (OuterVolumeSpecName: "scripts") pod "98c79518-5ef3-4382-a156-6de22fc92082" (UID: "98c79518-5ef3-4382-a156-6de22fc92082"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.759878 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98c79518-5ef3-4382-a156-6de22fc92082" (UID: "98c79518-5ef3-4382-a156-6de22fc92082"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.763184 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "98c79518-5ef3-4382-a156-6de22fc92082" (UID: "98c79518-5ef3-4382-a156-6de22fc92082"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.831816 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.832009 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfx9c\" (UniqueName: \"kubernetes.io/projected/98c79518-5ef3-4382-a156-6de22fc92082-kube-api-access-nfx9c\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.832062 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvl87\" (UniqueName: \"kubernetes.io/projected/91f8d060-b713-4586-b388-c68c51647297-kube-api-access-zvl87\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.832151 4809 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98c79518-5ef3-4382-a156-6de22fc92082-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.832203 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98c79518-5ef3-4382-a156-6de22fc92082-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.832254 4809 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.832321 4809 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98c79518-5ef3-4382-a156-6de22fc92082-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:12 crc kubenswrapper[4809]: I1127 17:38:12.832378 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2slq\" (UniqueName: \"kubernetes.io/projected/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f-kube-api-access-r2slq\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.128209 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ql6vt" event={"ID":"16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f","Type":"ContainerDied","Data":"dc46d8df8b54a853fd313727feed3e4d80d0a9000f281a70c350424c30b8ea6c"} Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.128567 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc46d8df8b54a853fd313727feed3e4d80d0a9000f281a70c350424c30b8ea6c" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.128241 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ql6vt" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.130554 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-54b8-account-create-update-4t2zg" event={"ID":"91f8d060-b713-4586-b388-c68c51647297","Type":"ContainerDied","Data":"7f3e1be2dc9f2596a5e0005a1474caf6f4def899e24974ae78904347b3a31997"} Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.130596 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f3e1be2dc9f2596a5e0005a1474caf6f4def899e24974ae78904347b3a31997" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.130657 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-54b8-account-create-update-4t2zg" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.132847 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hgrfv" event={"ID":"98c79518-5ef3-4382-a156-6de22fc92082","Type":"ContainerDied","Data":"1a1c4f63d4595f9ee9c813340224a79d1ef38fea369dd7090a650a60aef1ad68"} Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.132892 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a1c4f63d4595f9ee9c813340224a79d1ef38fea369dd7090a650a60aef1ad68" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.132913 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hgrfv" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.242391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.262908 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1b359048-c31b-4d16-922a-77b4b9afe87d-etc-swift\") pod \"swift-storage-0\" (UID: \"1b359048-c31b-4d16-922a-77b4b9afe87d\") " pod="openstack/swift-storage-0" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.507058 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.824622 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-ldz52"] Nov 27 17:38:13 crc kubenswrapper[4809]: E1127 17:38:13.825354 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f8d060-b713-4586-b388-c68c51647297" containerName="mariadb-account-create-update" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.825373 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f8d060-b713-4586-b388-c68c51647297" containerName="mariadb-account-create-update" Nov 27 17:38:13 crc kubenswrapper[4809]: E1127 17:38:13.825387 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f" containerName="mariadb-database-create" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.825395 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f" containerName="mariadb-database-create" Nov 27 17:38:13 crc kubenswrapper[4809]: E1127 17:38:13.825416 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e56ebf-c6e1-4675-afcb-cbb600ac778e" containerName="dnsmasq-dns" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.825425 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e56ebf-c6e1-4675-afcb-cbb600ac778e" containerName="dnsmasq-dns" Nov 27 17:38:13 crc kubenswrapper[4809]: E1127 17:38:13.825442 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c79518-5ef3-4382-a156-6de22fc92082" containerName="swift-ring-rebalance" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.825449 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c79518-5ef3-4382-a156-6de22fc92082" containerName="swift-ring-rebalance" Nov 27 17:38:13 crc kubenswrapper[4809]: E1127 17:38:13.825464 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e56ebf-c6e1-4675-afcb-cbb600ac778e" containerName="init" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.825472 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e56ebf-c6e1-4675-afcb-cbb600ac778e" containerName="init" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.825666 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c79518-5ef3-4382-a156-6de22fc92082" containerName="swift-ring-rebalance" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.825686 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f" containerName="mariadb-database-create" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.825698 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e56ebf-c6e1-4675-afcb-cbb600ac778e" containerName="dnsmasq-dns" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.825729 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f8d060-b713-4586-b388-c68c51647297" containerName="mariadb-account-create-update" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.826476 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.836791 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e5a0-account-create-update-px7dm"] Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.838169 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.840828 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.846157 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ldz52"] Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.853019 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pstz\" (UniqueName: \"kubernetes.io/projected/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-kube-api-access-7pstz\") pod \"keystone-e5a0-account-create-update-px7dm\" (UID: \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\") " pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.853257 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-operator-scripts\") pod \"keystone-e5a0-account-create-update-px7dm\" (UID: \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\") " pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.886691 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e5a0-account-create-update-px7dm"] Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.954321 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9da48a67-65c8-462f-91af-ed6c4192010d-operator-scripts\") pod \"keystone-db-create-ldz52\" (UID: \"9da48a67-65c8-462f-91af-ed6c4192010d\") " pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.955046 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-operator-scripts\") pod \"keystone-e5a0-account-create-update-px7dm\" (UID: \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\") " pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.955167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pstz\" (UniqueName: \"kubernetes.io/projected/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-kube-api-access-7pstz\") pod \"keystone-e5a0-account-create-update-px7dm\" (UID: \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\") " pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.955241 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xg9q\" (UniqueName: \"kubernetes.io/projected/9da48a67-65c8-462f-91af-ed6c4192010d-kube-api-access-5xg9q\") pod \"keystone-db-create-ldz52\" (UID: \"9da48a67-65c8-462f-91af-ed6c4192010d\") " pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.955630 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-operator-scripts\") pod \"keystone-e5a0-account-create-update-px7dm\" (UID: \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\") " pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:13 crc kubenswrapper[4809]: I1127 17:38:13.975604 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pstz\" (UniqueName: \"kubernetes.io/projected/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-kube-api-access-7pstz\") pod \"keystone-e5a0-account-create-update-px7dm\" (UID: \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\") " pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.026872 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 27 17:38:14 crc kubenswrapper[4809]: W1127 17:38:14.029906 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b359048_c31b_4d16_922a_77b4b9afe87d.slice/crio-ba574e83dd21b993b9ca8aeee89be9b9b9f5bc40af7f308c57ccc2eeb2d7f347 WatchSource:0}: Error finding container ba574e83dd21b993b9ca8aeee89be9b9b9f5bc40af7f308c57ccc2eeb2d7f347: Status 404 returned error can't find the container with id ba574e83dd21b993b9ca8aeee89be9b9b9f5bc40af7f308c57ccc2eeb2d7f347 Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.056852 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xg9q\" (UniqueName: \"kubernetes.io/projected/9da48a67-65c8-462f-91af-ed6c4192010d-kube-api-access-5xg9q\") pod \"keystone-db-create-ldz52\" (UID: \"9da48a67-65c8-462f-91af-ed6c4192010d\") " pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.057037 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9da48a67-65c8-462f-91af-ed6c4192010d-operator-scripts\") pod \"keystone-db-create-ldz52\" (UID: \"9da48a67-65c8-462f-91af-ed6c4192010d\") " pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.057881 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9da48a67-65c8-462f-91af-ed6c4192010d-operator-scripts\") pod \"keystone-db-create-ldz52\" (UID: \"9da48a67-65c8-462f-91af-ed6c4192010d\") " pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.072285 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xg9q\" (UniqueName: \"kubernetes.io/projected/9da48a67-65c8-462f-91af-ed6c4192010d-kube-api-access-5xg9q\") pod \"keystone-db-create-ldz52\" (UID: \"9da48a67-65c8-462f-91af-ed6c4192010d\") " pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.096181 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-vqr6f"] Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.097462 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.114700 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vqr6f"] Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.142597 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"ba574e83dd21b993b9ca8aeee89be9b9b9f5bc40af7f308c57ccc2eeb2d7f347"} Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.155290 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.160944 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-295b6\" (UniqueName: \"kubernetes.io/projected/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-kube-api-access-295b6\") pod \"placement-db-create-vqr6f\" (UID: \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\") " pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.161059 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-operator-scripts\") pod \"placement-db-create-vqr6f\" (UID: \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\") " pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.169384 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.210864 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8e30-account-create-update-b8pvk"] Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.212521 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.217232 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.219278 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8e30-account-create-update-b8pvk"] Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.262688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-operator-scripts\") pod \"placement-db-create-vqr6f\" (UID: \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\") " pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.263154 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-295b6\" (UniqueName: \"kubernetes.io/projected/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-kube-api-access-295b6\") pod \"placement-db-create-vqr6f\" (UID: \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\") " pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.266224 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-operator-scripts\") pod \"placement-db-create-vqr6f\" (UID: \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\") " pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.289334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-295b6\" (UniqueName: \"kubernetes.io/projected/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-kube-api-access-295b6\") pod \"placement-db-create-vqr6f\" (UID: \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\") " pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.365358 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrw96\" (UniqueName: \"kubernetes.io/projected/61639b29-dc56-47d2-8c8d-44aea5ba45f7-kube-api-access-jrw96\") pod \"placement-8e30-account-create-update-b8pvk\" (UID: \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\") " pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.365441 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61639b29-dc56-47d2-8c8d-44aea5ba45f7-operator-scripts\") pod \"placement-8e30-account-create-update-b8pvk\" (UID: \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\") " pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.428544 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.467128 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrw96\" (UniqueName: \"kubernetes.io/projected/61639b29-dc56-47d2-8c8d-44aea5ba45f7-kube-api-access-jrw96\") pod \"placement-8e30-account-create-update-b8pvk\" (UID: \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\") " pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.467195 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61639b29-dc56-47d2-8c8d-44aea5ba45f7-operator-scripts\") pod \"placement-8e30-account-create-update-b8pvk\" (UID: \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\") " pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.467911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61639b29-dc56-47d2-8c8d-44aea5ba45f7-operator-scripts\") pod \"placement-8e30-account-create-update-b8pvk\" (UID: \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\") " pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.483887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrw96\" (UniqueName: \"kubernetes.io/projected/61639b29-dc56-47d2-8c8d-44aea5ba45f7-kube-api-access-jrw96\") pod \"placement-8e30-account-create-update-b8pvk\" (UID: \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\") " pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.532416 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.535909 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-bm4mw"] Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.537021 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.543160 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.543369 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bh9xs" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.548772 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-bm4mw"] Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.642425 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ldz52"] Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.683081 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fcwg\" (UniqueName: \"kubernetes.io/projected/acc5484a-4058-4a5d-97a0-e84aaa75905f-kube-api-access-9fcwg\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.683311 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-db-sync-config-data\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.683484 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-config-data\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.683629 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-combined-ca-bundle\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.707010 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e5a0-account-create-update-px7dm"] Nov 27 17:38:14 crc kubenswrapper[4809]: W1127 17:38:14.715371 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a54a5ce_419a_4b39_9fc9_cd80dd4c1e61.slice/crio-9318da2d7f1e6b9301b80876bdddd6bdaff47404e203d73bba459a3d8745e015 WatchSource:0}: Error finding container 9318da2d7f1e6b9301b80876bdddd6bdaff47404e203d73bba459a3d8745e015: Status 404 returned error can't find the container with id 9318da2d7f1e6b9301b80876bdddd6bdaff47404e203d73bba459a3d8745e015 Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.785904 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-db-sync-config-data\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.786012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-config-data\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.786110 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-combined-ca-bundle\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.786149 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fcwg\" (UniqueName: \"kubernetes.io/projected/acc5484a-4058-4a5d-97a0-e84aaa75905f-kube-api-access-9fcwg\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.794474 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-config-data\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.797329 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-db-sync-config-data\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.799825 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-combined-ca-bundle\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.801976 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fcwg\" (UniqueName: \"kubernetes.io/projected/acc5484a-4058-4a5d-97a0-e84aaa75905f-kube-api-access-9fcwg\") pod \"glance-db-sync-bm4mw\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:14 crc kubenswrapper[4809]: I1127 17:38:14.988996 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vqr6f"] Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.006368 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.082560 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8e30-account-create-update-b8pvk"] Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.167047 4809 generic.go:334] "Generic (PLEG): container finished" podID="9da48a67-65c8-462f-91af-ed6c4192010d" containerID="f61380d9bd12c2bcdd88414cdbf980b53d492a95de6531d04ec443c17f571c72" exitCode=0 Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.167397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ldz52" event={"ID":"9da48a67-65c8-462f-91af-ed6c4192010d","Type":"ContainerDied","Data":"f61380d9bd12c2bcdd88414cdbf980b53d492a95de6531d04ec443c17f571c72"} Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.167496 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ldz52" event={"ID":"9da48a67-65c8-462f-91af-ed6c4192010d","Type":"ContainerStarted","Data":"99056f1681007529c1b83779d39af34b40f80a1731157ade0614f6e32d9669e5"} Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.168709 4809 generic.go:334] "Generic (PLEG): container finished" podID="2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61" containerID="76b55033f40a72810e20207f7f86639cf89daa9f50675468672928fb92f8d130" exitCode=0 Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.168920 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e5a0-account-create-update-px7dm" event={"ID":"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61","Type":"ContainerDied","Data":"76b55033f40a72810e20207f7f86639cf89daa9f50675468672928fb92f8d130"} Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.169010 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e5a0-account-create-update-px7dm" event={"ID":"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61","Type":"ContainerStarted","Data":"9318da2d7f1e6b9301b80876bdddd6bdaff47404e203d73bba459a3d8745e015"} Nov 27 17:38:15 crc kubenswrapper[4809]: W1127 17:38:15.174401 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b9207d8_29be_4fa6_bd67_04eb1afc29a7.slice/crio-0f35680bea05290b9c1792282c75280e1eccc3596e604d4352c164c1aaf257fb WatchSource:0}: Error finding container 0f35680bea05290b9c1792282c75280e1eccc3596e604d4352c164c1aaf257fb: Status 404 returned error can't find the container with id 0f35680bea05290b9c1792282c75280e1eccc3596e604d4352c164c1aaf257fb Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.230541 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-h8g6t" podUID="55d50ebb-8cab-42df-96b5-9598262337a4" containerName="ovn-controller" probeResult="failure" output=< Nov 27 17:38:15 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 27 17:38:15 crc kubenswrapper[4809]: > Nov 27 17:38:15 crc kubenswrapper[4809]: I1127 17:38:15.745495 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-bm4mw"] Nov 27 17:38:15 crc kubenswrapper[4809]: W1127 17:38:15.755754 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacc5484a_4058_4a5d_97a0_e84aaa75905f.slice/crio-646c934206be17400d57ed844c6cddb68cdf6fe2a393cf3143772c283434eeb1 WatchSource:0}: Error finding container 646c934206be17400d57ed844c6cddb68cdf6fe2a393cf3143772c283434eeb1: Status 404 returned error can't find the container with id 646c934206be17400d57ed844c6cddb68cdf6fe2a393cf3143772c283434eeb1 Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.184646 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"611decf12f8238e22dafb3d9916e9e18bad7e52d7346d3d524f1d2a1ddb7c0e2"} Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.185109 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"be1dcc0824ea3b60a0d53eaea9b6963f875454f586ce1a6911db1e4ee3772a45"} Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.185120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"7e3cdb655c35cf4fcf4c4517992bc734f813e42d692c10caa4867b1f159fd82a"} Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.187663 4809 generic.go:334] "Generic (PLEG): container finished" podID="5b9207d8-29be-4fa6-bd67-04eb1afc29a7" containerID="300d34c50c0b5b59ec864c21ee6e262af11ca7a6702fb6464dccf815090c3bf4" exitCode=0 Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.187726 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vqr6f" event={"ID":"5b9207d8-29be-4fa6-bd67-04eb1afc29a7","Type":"ContainerDied","Data":"300d34c50c0b5b59ec864c21ee6e262af11ca7a6702fb6464dccf815090c3bf4"} Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.187768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vqr6f" event={"ID":"5b9207d8-29be-4fa6-bd67-04eb1afc29a7","Type":"ContainerStarted","Data":"0f35680bea05290b9c1792282c75280e1eccc3596e604d4352c164c1aaf257fb"} Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.190507 4809 generic.go:334] "Generic (PLEG): container finished" podID="61639b29-dc56-47d2-8c8d-44aea5ba45f7" containerID="aba6cb2b85c552c5ad546aeb8dba03d515379e238c98345f5309dfd8081cca65" exitCode=0 Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.190644 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8e30-account-create-update-b8pvk" event={"ID":"61639b29-dc56-47d2-8c8d-44aea5ba45f7","Type":"ContainerDied","Data":"aba6cb2b85c552c5ad546aeb8dba03d515379e238c98345f5309dfd8081cca65"} Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.190721 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8e30-account-create-update-b8pvk" event={"ID":"61639b29-dc56-47d2-8c8d-44aea5ba45f7","Type":"ContainerStarted","Data":"eb964ad77226656e119bd153775ea3be4f7bac9cdf8089606a86730b7515f8f7"} Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.191970 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bm4mw" event={"ID":"acc5484a-4058-4a5d-97a0-e84aaa75905f","Type":"ContainerStarted","Data":"646c934206be17400d57ed844c6cddb68cdf6fe2a393cf3143772c283434eeb1"} Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.790602 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.795860 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.926890 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pstz\" (UniqueName: \"kubernetes.io/projected/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-kube-api-access-7pstz\") pod \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\" (UID: \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\") " Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.926938 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xg9q\" (UniqueName: \"kubernetes.io/projected/9da48a67-65c8-462f-91af-ed6c4192010d-kube-api-access-5xg9q\") pod \"9da48a67-65c8-462f-91af-ed6c4192010d\" (UID: \"9da48a67-65c8-462f-91af-ed6c4192010d\") " Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.926990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9da48a67-65c8-462f-91af-ed6c4192010d-operator-scripts\") pod \"9da48a67-65c8-462f-91af-ed6c4192010d\" (UID: \"9da48a67-65c8-462f-91af-ed6c4192010d\") " Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.927095 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-operator-scripts\") pod \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\" (UID: \"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61\") " Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.929048 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61" (UID: "2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.929488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9da48a67-65c8-462f-91af-ed6c4192010d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9da48a67-65c8-462f-91af-ed6c4192010d" (UID: "9da48a67-65c8-462f-91af-ed6c4192010d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.934920 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-kube-api-access-7pstz" (OuterVolumeSpecName: "kube-api-access-7pstz") pod "2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61" (UID: "2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61"). InnerVolumeSpecName "kube-api-access-7pstz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:16 crc kubenswrapper[4809]: I1127 17:38:16.935308 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9da48a67-65c8-462f-91af-ed6c4192010d-kube-api-access-5xg9q" (OuterVolumeSpecName: "kube-api-access-5xg9q") pod "9da48a67-65c8-462f-91af-ed6c4192010d" (UID: "9da48a67-65c8-462f-91af-ed6c4192010d"). InnerVolumeSpecName "kube-api-access-5xg9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.039236 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pstz\" (UniqueName: \"kubernetes.io/projected/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-kube-api-access-7pstz\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.039279 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xg9q\" (UniqueName: \"kubernetes.io/projected/9da48a67-65c8-462f-91af-ed6c4192010d-kube-api-access-5xg9q\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.039292 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9da48a67-65c8-462f-91af-ed6c4192010d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.039305 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.204353 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ldz52" event={"ID":"9da48a67-65c8-462f-91af-ed6c4192010d","Type":"ContainerDied","Data":"99056f1681007529c1b83779d39af34b40f80a1731157ade0614f6e32d9669e5"} Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.204384 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ldz52" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.204393 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99056f1681007529c1b83779d39af34b40f80a1731157ade0614f6e32d9669e5" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.208526 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"1b242a9483fefdca37c3d285cbd59a62584568abfa30703834a5f0978693c44c"} Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.211114 4809 generic.go:334] "Generic (PLEG): container finished" podID="f8145e23-a84e-405f-beb6-e27bb2aa1c59" containerID="e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5" exitCode=0 Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.211186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8145e23-a84e-405f-beb6-e27bb2aa1c59","Type":"ContainerDied","Data":"e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5"} Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.213125 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e5a0-account-create-update-px7dm" event={"ID":"2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61","Type":"ContainerDied","Data":"9318da2d7f1e6b9301b80876bdddd6bdaff47404e203d73bba459a3d8745e015"} Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.213154 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9318da2d7f1e6b9301b80876bdddd6bdaff47404e203d73bba459a3d8745e015" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.213164 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e5a0-account-create-update-px7dm" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.219131 4809 generic.go:334] "Generic (PLEG): container finished" podID="01288244-f670-47a4-ac43-180191d7f331" containerID="dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b" exitCode=0 Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.219216 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01288244-f670-47a4-ac43-180191d7f331","Type":"ContainerDied","Data":"dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b"} Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.865995 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.895889 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.959190 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrw96\" (UniqueName: \"kubernetes.io/projected/61639b29-dc56-47d2-8c8d-44aea5ba45f7-kube-api-access-jrw96\") pod \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\" (UID: \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\") " Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.960004 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61639b29-dc56-47d2-8c8d-44aea5ba45f7-operator-scripts\") pod \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\" (UID: \"61639b29-dc56-47d2-8c8d-44aea5ba45f7\") " Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.960882 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61639b29-dc56-47d2-8c8d-44aea5ba45f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "61639b29-dc56-47d2-8c8d-44aea5ba45f7" (UID: "61639b29-dc56-47d2-8c8d-44aea5ba45f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:17 crc kubenswrapper[4809]: I1127 17:38:17.966264 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61639b29-dc56-47d2-8c8d-44aea5ba45f7-kube-api-access-jrw96" (OuterVolumeSpecName: "kube-api-access-jrw96") pod "61639b29-dc56-47d2-8c8d-44aea5ba45f7" (UID: "61639b29-dc56-47d2-8c8d-44aea5ba45f7"). InnerVolumeSpecName "kube-api-access-jrw96". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.061677 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-295b6\" (UniqueName: \"kubernetes.io/projected/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-kube-api-access-295b6\") pod \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\" (UID: \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\") " Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.062650 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-operator-scripts\") pod \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\" (UID: \"5b9207d8-29be-4fa6-bd67-04eb1afc29a7\") " Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.063216 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5b9207d8-29be-4fa6-bd67-04eb1afc29a7" (UID: "5b9207d8-29be-4fa6-bd67-04eb1afc29a7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.063647 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61639b29-dc56-47d2-8c8d-44aea5ba45f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.063669 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrw96\" (UniqueName: \"kubernetes.io/projected/61639b29-dc56-47d2-8c8d-44aea5ba45f7-kube-api-access-jrw96\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.063678 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.066290 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-kube-api-access-295b6" (OuterVolumeSpecName: "kube-api-access-295b6") pod "5b9207d8-29be-4fa6-bd67-04eb1afc29a7" (UID: "5b9207d8-29be-4fa6-bd67-04eb1afc29a7"). InnerVolumeSpecName "kube-api-access-295b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.167867 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-295b6\" (UniqueName: \"kubernetes.io/projected/5b9207d8-29be-4fa6-bd67-04eb1afc29a7-kube-api-access-295b6\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.233070 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"4be1d41677bd9495aa5a5c114749160835e14cbd75623e26a653f3fe65f7cda2"} Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.233133 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"c64c847b8e1e6148eab43249aed85521e45648858404d6d1d348ca83dfc15b66"} Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.241061 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8145e23-a84e-405f-beb6-e27bb2aa1c59","Type":"ContainerStarted","Data":"9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d"} Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.243190 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.245135 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vqr6f" event={"ID":"5b9207d8-29be-4fa6-bd67-04eb1afc29a7","Type":"ContainerDied","Data":"0f35680bea05290b9c1792282c75280e1eccc3596e604d4352c164c1aaf257fb"} Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.245165 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f35680bea05290b9c1792282c75280e1eccc3596e604d4352c164c1aaf257fb" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.245213 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vqr6f" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.257437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01288244-f670-47a4-ac43-180191d7f331","Type":"ContainerStarted","Data":"5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439"} Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.259698 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.263766 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8e30-account-create-update-b8pvk" event={"ID":"61639b29-dc56-47d2-8c8d-44aea5ba45f7","Type":"ContainerDied","Data":"eb964ad77226656e119bd153775ea3be4f7bac9cdf8089606a86730b7515f8f7"} Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.263796 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb964ad77226656e119bd153775ea3be4f7bac9cdf8089606a86730b7515f8f7" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.263839 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8e30-account-create-update-b8pvk" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.272335 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.345294703 podStartE2EDuration="59.272317169s" podCreationTimestamp="2025-11-27 17:37:19 +0000 UTC" firstStartedPulling="2025-11-27 17:37:33.466267461 +0000 UTC m=+1688.738724813" lastFinishedPulling="2025-11-27 17:37:42.393289937 +0000 UTC m=+1697.665747279" observedRunningTime="2025-11-27 17:38:18.268297389 +0000 UTC m=+1733.540754741" watchObservedRunningTime="2025-11-27 17:38:18.272317169 +0000 UTC m=+1733.544774521" Nov 27 17:38:18 crc kubenswrapper[4809]: I1127 17:38:18.320840 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.422995717 podStartE2EDuration="59.320816928s" podCreationTimestamp="2025-11-27 17:37:19 +0000 UTC" firstStartedPulling="2025-11-27 17:37:33.495467506 +0000 UTC m=+1688.767924858" lastFinishedPulling="2025-11-27 17:37:42.393288717 +0000 UTC m=+1697.665746069" observedRunningTime="2025-11-27 17:38:18.314579339 +0000 UTC m=+1733.587036691" watchObservedRunningTime="2025-11-27 17:38:18.320816928 +0000 UTC m=+1733.593274280" Nov 27 17:38:19 crc kubenswrapper[4809]: I1127 17:38:19.004139 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 27 17:38:19 crc kubenswrapper[4809]: I1127 17:38:19.280763 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"60bba75e3ab64195f158b785bd850818087a47327fc98d1930915bd7d1769e05"} Nov 27 17:38:19 crc kubenswrapper[4809]: I1127 17:38:19.281179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"7e3e150f4d0754228414219c36c59a91c7b3bbd6df109a51833594436270e514"} Nov 27 17:38:20 crc kubenswrapper[4809]: I1127 17:38:20.243988 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-h8g6t" podUID="55d50ebb-8cab-42df-96b5-9598262337a4" containerName="ovn-controller" probeResult="failure" output=< Nov 27 17:38:20 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 27 17:38:20 crc kubenswrapper[4809]: > Nov 27 17:38:20 crc kubenswrapper[4809]: I1127 17:38:20.280843 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:38:21 crc kubenswrapper[4809]: I1127 17:38:21.306363 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"c00956b6b46a3bcc4df35f5d7b2113d2222f88f8862ed95480cfff4c5650d147"} Nov 27 17:38:21 crc kubenswrapper[4809]: I1127 17:38:21.306989 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"422bb033688d8b5aa5ae24979f72be34e2bcabc61369d7df59399612fc95638f"} Nov 27 17:38:22 crc kubenswrapper[4809]: I1127 17:38:22.325309 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"7a514f5c93e917c9cabc7132918d9a49b9c4e2285ce66e1ea8561063f9be2e19"} Nov 27 17:38:22 crc kubenswrapper[4809]: I1127 17:38:22.325622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"84d526657034c46121b25744acfd9a58860748902d11387a94a59655fc6d0dc3"} Nov 27 17:38:22 crc kubenswrapper[4809]: I1127 17:38:22.325632 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"53d9f5132c927b5ee6725f88b9764fa936042687cb5715176b850460d0787f81"} Nov 27 17:38:22 crc kubenswrapper[4809]: I1127 17:38:22.325640 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"6aa30d7b04840706698dba9ee6b74be482220f8b51ac9307921316a1a0e51d98"} Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.244339 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-h8g6t" podUID="55d50ebb-8cab-42df-96b5-9598262337a4" containerName="ovn-controller" probeResult="failure" output=< Nov 27 17:38:25 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 27 17:38:25 crc kubenswrapper[4809]: > Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.287881 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jnvf9" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.524717 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-h8g6t-config-rhl6b"] Nov 27 17:38:25 crc kubenswrapper[4809]: E1127 17:38:25.525447 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61639b29-dc56-47d2-8c8d-44aea5ba45f7" containerName="mariadb-account-create-update" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.525513 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="61639b29-dc56-47d2-8c8d-44aea5ba45f7" containerName="mariadb-account-create-update" Nov 27 17:38:25 crc kubenswrapper[4809]: E1127 17:38:25.525608 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9da48a67-65c8-462f-91af-ed6c4192010d" containerName="mariadb-database-create" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.525667 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9da48a67-65c8-462f-91af-ed6c4192010d" containerName="mariadb-database-create" Nov 27 17:38:25 crc kubenswrapper[4809]: E1127 17:38:25.525728 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61" containerName="mariadb-account-create-update" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.525826 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61" containerName="mariadb-account-create-update" Nov 27 17:38:25 crc kubenswrapper[4809]: E1127 17:38:25.525885 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9207d8-29be-4fa6-bd67-04eb1afc29a7" containerName="mariadb-database-create" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.525935 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9207d8-29be-4fa6-bd67-04eb1afc29a7" containerName="mariadb-database-create" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.526152 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9207d8-29be-4fa6-bd67-04eb1afc29a7" containerName="mariadb-database-create" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.526213 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="61639b29-dc56-47d2-8c8d-44aea5ba45f7" containerName="mariadb-account-create-update" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.526266 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9da48a67-65c8-462f-91af-ed6c4192010d" containerName="mariadb-database-create" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.526327 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61" containerName="mariadb-account-create-update" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.527019 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.529233 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.534765 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h8g6t-config-rhl6b"] Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.609866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-scripts\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.609929 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run-ovn\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.609990 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.610336 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-additional-scripts\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.610384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-log-ovn\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.610426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2crfc\" (UniqueName: \"kubernetes.io/projected/1ae0eae0-702d-485d-8438-db6f005ad77f-kube-api-access-2crfc\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.712195 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-additional-scripts\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.712453 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-log-ovn\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.712481 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2crfc\" (UniqueName: \"kubernetes.io/projected/1ae0eae0-702d-485d-8438-db6f005ad77f-kube-api-access-2crfc\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.712525 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-scripts\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.712544 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run-ovn\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.712584 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.712911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.712969 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run-ovn\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.713017 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-additional-scripts\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.713257 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-log-ovn\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.715313 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-scripts\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.734029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2crfc\" (UniqueName: \"kubernetes.io/projected/1ae0eae0-702d-485d-8438-db6f005ad77f-kube-api-access-2crfc\") pod \"ovn-controller-h8g6t-config-rhl6b\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:25 crc kubenswrapper[4809]: I1127 17:38:25.848369 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:26 crc kubenswrapper[4809]: I1127 17:38:26.458323 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:38:26 crc kubenswrapper[4809]: E1127 17:38:26.458558 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:38:29 crc kubenswrapper[4809]: I1127 17:38:29.867504 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h8g6t-config-rhl6b"] Nov 27 17:38:29 crc kubenswrapper[4809]: W1127 17:38:29.876933 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ae0eae0_702d_485d_8438_db6f005ad77f.slice/crio-8ba5525c062f8785ffcc1ad6d04804e42842127e437994e0791e7f67ce2eaf74 WatchSource:0}: Error finding container 8ba5525c062f8785ffcc1ad6d04804e42842127e437994e0791e7f67ce2eaf74: Status 404 returned error can't find the container with id 8ba5525c062f8785ffcc1ad6d04804e42842127e437994e0791e7f67ce2eaf74 Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.222834 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-h8g6t" podUID="55d50ebb-8cab-42df-96b5-9598262337a4" containerName="ovn-controller" probeResult="failure" output=< Nov 27 17:38:30 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 27 17:38:30 crc kubenswrapper[4809]: > Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.410464 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t-config-rhl6b" event={"ID":"1ae0eae0-702d-485d-8438-db6f005ad77f","Type":"ContainerStarted","Data":"641fe1a719780307288b3c357021f5dd62ba4d1d0e1af01d4cef3bb1bf38dff5"} Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.410805 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t-config-rhl6b" event={"ID":"1ae0eae0-702d-485d-8438-db6f005ad77f","Type":"ContainerStarted","Data":"8ba5525c062f8785ffcc1ad6d04804e42842127e437994e0791e7f67ce2eaf74"} Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.412864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bm4mw" event={"ID":"acc5484a-4058-4a5d-97a0-e84aaa75905f","Type":"ContainerStarted","Data":"b52055a16d85e6af2503992f7364d1eb40fb9f67870308ae59d2dc63673b2fd0"} Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.417854 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1b359048-c31b-4d16-922a-77b4b9afe87d","Type":"ContainerStarted","Data":"50974da23f63ab0ac5a50ad1e3a8212e209b759a9a362d7418a053b8f7fb03f9"} Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.485580 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=28.287943653 podStartE2EDuration="34.485553139s" podCreationTimestamp="2025-11-27 17:37:56 +0000 UTC" firstStartedPulling="2025-11-27 17:38:14.031758299 +0000 UTC m=+1729.304215651" lastFinishedPulling="2025-11-27 17:38:20.229367785 +0000 UTC m=+1735.501825137" observedRunningTime="2025-11-27 17:38:30.477651285 +0000 UTC m=+1745.750108647" watchObservedRunningTime="2025-11-27 17:38:30.485553139 +0000 UTC m=+1745.758010491" Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.499910 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.500253 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-bm4mw" podStartSLOduration=2.798466253 podStartE2EDuration="16.500231179s" podCreationTimestamp="2025-11-27 17:38:14 +0000 UTC" firstStartedPulling="2025-11-27 17:38:15.761753216 +0000 UTC m=+1731.034210568" lastFinishedPulling="2025-11-27 17:38:29.463518142 +0000 UTC m=+1744.735975494" observedRunningTime="2025-11-27 17:38:30.492840178 +0000 UTC m=+1745.765297530" watchObservedRunningTime="2025-11-27 17:38:30.500231179 +0000 UTC m=+1745.772688541" Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.821918 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.930808 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-w5bp6"] Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.932319 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.940074 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 27 17:38:30 crc kubenswrapper[4809]: I1127 17:38:30.979519 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-w5bp6"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.020034 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.020147 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.020184 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qkmx\" (UniqueName: \"kubernetes.io/projected/d542ca5d-c018-4f64-96f0-e515d148c51c-kube-api-access-5qkmx\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.020218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.020287 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.020334 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-config\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.082671 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vg6fv"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.093281 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.100335 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vg6fv"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.130542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.130609 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.130635 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qkmx\" (UniqueName: \"kubernetes.io/projected/d542ca5d-c018-4f64-96f0-e515d148c51c-kube-api-access-5qkmx\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.130661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.130691 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.130721 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-config\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.131474 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-config\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.132149 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.132367 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.132874 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.133228 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.186601 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qkmx\" (UniqueName: \"kubernetes.io/projected/d542ca5d-c018-4f64-96f0-e515d148c51c-kube-api-access-5qkmx\") pod \"dnsmasq-dns-77585f5f8c-w5bp6\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.209393 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-8bnvg"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.210669 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.216838 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-87f3-account-create-update-klgkz"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.218178 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.221703 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.233699 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr6vb\" (UniqueName: \"kubernetes.io/projected/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-kube-api-access-cr6vb\") pod \"cinder-db-create-vg6fv\" (UID: \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\") " pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.233829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-operator-scripts\") pod \"cinder-db-create-vg6fv\" (UID: \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\") " pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.236108 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8bnvg"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.250707 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-87f3-account-create-update-klgkz"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.262831 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.318327 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-vtslx"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.319706 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.337904 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b7qp\" (UniqueName: \"kubernetes.io/projected/b374716e-7e54-4be1-96bf-3c9e9bd379b8-kube-api-access-4b7qp\") pod \"barbican-db-create-8bnvg\" (UID: \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\") " pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.337970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-operator-scripts\") pod \"cinder-db-create-vg6fv\" (UID: \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\") " pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.338027 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b374716e-7e54-4be1-96bf-3c9e9bd379b8-operator-scripts\") pod \"barbican-db-create-8bnvg\" (UID: \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\") " pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.338047 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wghhn\" (UniqueName: \"kubernetes.io/projected/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-kube-api-access-wghhn\") pod \"barbican-87f3-account-create-update-klgkz\" (UID: \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\") " pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.338076 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-operator-scripts\") pod \"barbican-87f3-account-create-update-klgkz\" (UID: \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\") " pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.338112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr6vb\" (UniqueName: \"kubernetes.io/projected/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-kube-api-access-cr6vb\") pod \"cinder-db-create-vg6fv\" (UID: \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\") " pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.339304 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-operator-scripts\") pod \"cinder-db-create-vg6fv\" (UID: \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\") " pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.355358 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-7fd4-account-create-update-pg2kx"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.356653 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.371271 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.381093 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-vtslx"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.391546 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7fd4-account-create-update-pg2kx"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.393036 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr6vb\" (UniqueName: \"kubernetes.io/projected/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-kube-api-access-cr6vb\") pod \"cinder-db-create-vg6fv\" (UID: \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\") " pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.400114 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-5tlhg"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.401595 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.405838 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.406076 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.406190 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.406404 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-l4mwz" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.426224 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-5tlhg"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.431151 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.440918 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b374716e-7e54-4be1-96bf-3c9e9bd379b8-operator-scripts\") pod \"barbican-db-create-8bnvg\" (UID: \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\") " pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.440953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wghhn\" (UniqueName: \"kubernetes.io/projected/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-kube-api-access-wghhn\") pod \"barbican-87f3-account-create-update-klgkz\" (UID: \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\") " pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.440981 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-operator-scripts\") pod \"barbican-87f3-account-create-update-klgkz\" (UID: \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\") " pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.441015 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcd2971-5635-42ab-af5c-ee867a78bfa1-operator-scripts\") pod \"neutron-db-create-vtslx\" (UID: \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\") " pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.441090 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktbnr\" (UniqueName: \"kubernetes.io/projected/4dcd2971-5635-42ab-af5c-ee867a78bfa1-kube-api-access-ktbnr\") pod \"neutron-db-create-vtslx\" (UID: \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\") " pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.441120 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b7qp\" (UniqueName: \"kubernetes.io/projected/b374716e-7e54-4be1-96bf-3c9e9bd379b8-kube-api-access-4b7qp\") pod \"barbican-db-create-8bnvg\" (UID: \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\") " pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.442110 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b374716e-7e54-4be1-96bf-3c9e9bd379b8-operator-scripts\") pod \"barbican-db-create-8bnvg\" (UID: \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\") " pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.442247 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-operator-scripts\") pod \"barbican-87f3-account-create-update-klgkz\" (UID: \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\") " pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.469285 4809 generic.go:334] "Generic (PLEG): container finished" podID="1ae0eae0-702d-485d-8438-db6f005ad77f" containerID="641fe1a719780307288b3c357021f5dd62ba4d1d0e1af01d4cef3bb1bf38dff5" exitCode=0 Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.475234 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b7qp\" (UniqueName: \"kubernetes.io/projected/b374716e-7e54-4be1-96bf-3c9e9bd379b8-kube-api-access-4b7qp\") pod \"barbican-db-create-8bnvg\" (UID: \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\") " pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.481608 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wghhn\" (UniqueName: \"kubernetes.io/projected/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-kube-api-access-wghhn\") pod \"barbican-87f3-account-create-update-klgkz\" (UID: \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\") " pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.487097 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t-config-rhl6b" event={"ID":"1ae0eae0-702d-485d-8438-db6f005ad77f","Type":"ContainerDied","Data":"641fe1a719780307288b3c357021f5dd62ba4d1d0e1af01d4cef3bb1bf38dff5"} Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.487138 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-19b4-account-create-update-n9t2g"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.490325 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.494234 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.501016 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-19b4-account-create-update-n9t2g"] Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.543163 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjgd6\" (UniqueName: \"kubernetes.io/projected/4891ffad-a908-49ac-97f5-196ae7a2e919-kube-api-access-wjgd6\") pod \"keystone-db-sync-5tlhg\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.543342 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-config-data\") pod \"keystone-db-sync-5tlhg\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.543501 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcd2971-5635-42ab-af5c-ee867a78bfa1-operator-scripts\") pod \"neutron-db-create-vtslx\" (UID: \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\") " pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.543551 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-combined-ca-bundle\") pod \"keystone-db-sync-5tlhg\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.543575 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm5tg\" (UniqueName: \"kubernetes.io/projected/025088fd-c5dc-452e-aceb-c4a04cc09033-kube-api-access-jm5tg\") pod \"cinder-7fd4-account-create-update-pg2kx\" (UID: \"025088fd-c5dc-452e-aceb-c4a04cc09033\") " pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.543639 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/025088fd-c5dc-452e-aceb-c4a04cc09033-operator-scripts\") pod \"cinder-7fd4-account-create-update-pg2kx\" (UID: \"025088fd-c5dc-452e-aceb-c4a04cc09033\") " pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.543666 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktbnr\" (UniqueName: \"kubernetes.io/projected/4dcd2971-5635-42ab-af5c-ee867a78bfa1-kube-api-access-ktbnr\") pod \"neutron-db-create-vtslx\" (UID: \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\") " pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.544415 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcd2971-5635-42ab-af5c-ee867a78bfa1-operator-scripts\") pod \"neutron-db-create-vtslx\" (UID: \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\") " pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.553931 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.576924 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.578336 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktbnr\" (UniqueName: \"kubernetes.io/projected/4dcd2971-5635-42ab-af5c-ee867a78bfa1-kube-api-access-ktbnr\") pod \"neutron-db-create-vtslx\" (UID: \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\") " pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.650130 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lndj\" (UniqueName: \"kubernetes.io/projected/d44f63e5-7e83-4cd8-a488-0e88f7635938-kube-api-access-4lndj\") pod \"neutron-19b4-account-create-update-n9t2g\" (UID: \"d44f63e5-7e83-4cd8-a488-0e88f7635938\") " pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.650771 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-combined-ca-bundle\") pod \"keystone-db-sync-5tlhg\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.650834 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm5tg\" (UniqueName: \"kubernetes.io/projected/025088fd-c5dc-452e-aceb-c4a04cc09033-kube-api-access-jm5tg\") pod \"cinder-7fd4-account-create-update-pg2kx\" (UID: \"025088fd-c5dc-452e-aceb-c4a04cc09033\") " pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.651006 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/025088fd-c5dc-452e-aceb-c4a04cc09033-operator-scripts\") pod \"cinder-7fd4-account-create-update-pg2kx\" (UID: \"025088fd-c5dc-452e-aceb-c4a04cc09033\") " pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.651092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjgd6\" (UniqueName: \"kubernetes.io/projected/4891ffad-a908-49ac-97f5-196ae7a2e919-kube-api-access-wjgd6\") pod \"keystone-db-sync-5tlhg\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.651222 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-config-data\") pod \"keystone-db-sync-5tlhg\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.651256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d44f63e5-7e83-4cd8-a488-0e88f7635938-operator-scripts\") pod \"neutron-19b4-account-create-update-n9t2g\" (UID: \"d44f63e5-7e83-4cd8-a488-0e88f7635938\") " pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.654861 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/025088fd-c5dc-452e-aceb-c4a04cc09033-operator-scripts\") pod \"cinder-7fd4-account-create-update-pg2kx\" (UID: \"025088fd-c5dc-452e-aceb-c4a04cc09033\") " pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.658366 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-combined-ca-bundle\") pod \"keystone-db-sync-5tlhg\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.662028 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-config-data\") pod \"keystone-db-sync-5tlhg\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.679900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjgd6\" (UniqueName: \"kubernetes.io/projected/4891ffad-a908-49ac-97f5-196ae7a2e919-kube-api-access-wjgd6\") pod \"keystone-db-sync-5tlhg\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.692245 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm5tg\" (UniqueName: \"kubernetes.io/projected/025088fd-c5dc-452e-aceb-c4a04cc09033-kube-api-access-jm5tg\") pod \"cinder-7fd4-account-create-update-pg2kx\" (UID: \"025088fd-c5dc-452e-aceb-c4a04cc09033\") " pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.731622 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.745920 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.752973 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d44f63e5-7e83-4cd8-a488-0e88f7635938-operator-scripts\") pod \"neutron-19b4-account-create-update-n9t2g\" (UID: \"d44f63e5-7e83-4cd8-a488-0e88f7635938\") " pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.753124 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lndj\" (UniqueName: \"kubernetes.io/projected/d44f63e5-7e83-4cd8-a488-0e88f7635938-kube-api-access-4lndj\") pod \"neutron-19b4-account-create-update-n9t2g\" (UID: \"d44f63e5-7e83-4cd8-a488-0e88f7635938\") " pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.756348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d44f63e5-7e83-4cd8-a488-0e88f7635938-operator-scripts\") pod \"neutron-19b4-account-create-update-n9t2g\" (UID: \"d44f63e5-7e83-4cd8-a488-0e88f7635938\") " pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.794863 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lndj\" (UniqueName: \"kubernetes.io/projected/d44f63e5-7e83-4cd8-a488-0e88f7635938-kube-api-access-4lndj\") pod \"neutron-19b4-account-create-update-n9t2g\" (UID: \"d44f63e5-7e83-4cd8-a488-0e88f7635938\") " pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:31 crc kubenswrapper[4809]: I1127 17:38:31.857360 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.022863 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-w5bp6"] Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.037022 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.067344 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.151307 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vg6fv"] Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.165323 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run\") pod \"1ae0eae0-702d-485d-8438-db6f005ad77f\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.165495 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run-ovn\") pod \"1ae0eae0-702d-485d-8438-db6f005ad77f\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.165667 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-log-ovn\") pod \"1ae0eae0-702d-485d-8438-db6f005ad77f\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.166007 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-additional-scripts\") pod \"1ae0eae0-702d-485d-8438-db6f005ad77f\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.166121 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-scripts\") pod \"1ae0eae0-702d-485d-8438-db6f005ad77f\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.166213 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2crfc\" (UniqueName: \"kubernetes.io/projected/1ae0eae0-702d-485d-8438-db6f005ad77f-kube-api-access-2crfc\") pod \"1ae0eae0-702d-485d-8438-db6f005ad77f\" (UID: \"1ae0eae0-702d-485d-8438-db6f005ad77f\") " Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.167904 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1ae0eae0-702d-485d-8438-db6f005ad77f" (UID: "1ae0eae0-702d-485d-8438-db6f005ad77f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.167969 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run" (OuterVolumeSpecName: "var-run") pod "1ae0eae0-702d-485d-8438-db6f005ad77f" (UID: "1ae0eae0-702d-485d-8438-db6f005ad77f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.168053 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1ae0eae0-702d-485d-8438-db6f005ad77f" (UID: "1ae0eae0-702d-485d-8438-db6f005ad77f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.168666 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1ae0eae0-702d-485d-8438-db6f005ad77f" (UID: "1ae0eae0-702d-485d-8438-db6f005ad77f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.170568 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-scripts" (OuterVolumeSpecName: "scripts") pod "1ae0eae0-702d-485d-8438-db6f005ad77f" (UID: "1ae0eae0-702d-485d-8438-db6f005ad77f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.171840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ae0eae0-702d-485d-8438-db6f005ad77f-kube-api-access-2crfc" (OuterVolumeSpecName: "kube-api-access-2crfc") pod "1ae0eae0-702d-485d-8438-db6f005ad77f" (UID: "1ae0eae0-702d-485d-8438-db6f005ad77f"). InnerVolumeSpecName "kube-api-access-2crfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.269227 4809 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.269296 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae0eae0-702d-485d-8438-db6f005ad77f-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.269314 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2crfc\" (UniqueName: \"kubernetes.io/projected/1ae0eae0-702d-485d-8438-db6f005ad77f-kube-api-access-2crfc\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.269335 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.269349 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.269364 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ae0eae0-702d-485d-8438-db6f005ad77f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.312421 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-87f3-account-create-update-klgkz"] Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.319978 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8bnvg"] Nov 27 17:38:32 crc kubenswrapper[4809]: W1127 17:38:32.330280 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4a9f9d1_a45a_49a1_95bd_1681a1b220ff.slice/crio-9020cf03049accbea57df8cee91fa0bfad6372f78d713675a1fe35b861897263 WatchSource:0}: Error finding container 9020cf03049accbea57df8cee91fa0bfad6372f78d713675a1fe35b861897263: Status 404 returned error can't find the container with id 9020cf03049accbea57df8cee91fa0bfad6372f78d713675a1fe35b861897263 Nov 27 17:38:32 crc kubenswrapper[4809]: W1127 17:38:32.330769 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb374716e_7e54_4be1_96bf_3c9e9bd379b8.slice/crio-664dcadb1a5c8988f2e6c2758480e02cbb4f97208609859b013968d5685b13d4 WatchSource:0}: Error finding container 664dcadb1a5c8988f2e6c2758480e02cbb4f97208609859b013968d5685b13d4: Status 404 returned error can't find the container with id 664dcadb1a5c8988f2e6c2758480e02cbb4f97208609859b013968d5685b13d4 Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.490953 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-87f3-account-create-update-klgkz" event={"ID":"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff","Type":"ContainerStarted","Data":"9020cf03049accbea57df8cee91fa0bfad6372f78d713675a1fe35b861897263"} Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.498103 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-5tlhg"] Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.498722 4809 generic.go:334] "Generic (PLEG): container finished" podID="d542ca5d-c018-4f64-96f0-e515d148c51c" containerID="7096d9c06d0856f926b099a322c37c172a73cb0eb76d6a9845306d90c2143807" exitCode=0 Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.498885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" event={"ID":"d542ca5d-c018-4f64-96f0-e515d148c51c","Type":"ContainerDied","Data":"7096d9c06d0856f926b099a322c37c172a73cb0eb76d6a9845306d90c2143807"} Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.499316 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" event={"ID":"d542ca5d-c018-4f64-96f0-e515d148c51c","Type":"ContainerStarted","Data":"077f2d7ab492c78999dae23c165755bb7816e0d6f334e2e11175a86a3c0aee0f"} Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.502686 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vg6fv" event={"ID":"4ae75b87-925b-4b0c-9a59-1d59f534f9f2","Type":"ContainerStarted","Data":"2c83e01330814afae94a2e95302d070a959f576dc9b7f275f32f672d4d8dd57f"} Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.502751 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vg6fv" event={"ID":"4ae75b87-925b-4b0c-9a59-1d59f534f9f2","Type":"ContainerStarted","Data":"b5db5fa7585cfeddac410e783daadc37f302e1e8d585a10d2d803bf4776820c0"} Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.505067 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8bnvg" event={"ID":"b374716e-7e54-4be1-96bf-3c9e9bd379b8","Type":"ContainerStarted","Data":"664dcadb1a5c8988f2e6c2758480e02cbb4f97208609859b013968d5685b13d4"} Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.516007 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t-config-rhl6b" event={"ID":"1ae0eae0-702d-485d-8438-db6f005ad77f","Type":"ContainerDied","Data":"8ba5525c062f8785ffcc1ad6d04804e42842127e437994e0791e7f67ce2eaf74"} Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.516041 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ba5525c062f8785ffcc1ad6d04804e42842127e437994e0791e7f67ce2eaf74" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.516105 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t-config-rhl6b" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.566386 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-vg6fv" podStartSLOduration=1.566365535 podStartE2EDuration="1.566365535s" podCreationTimestamp="2025-11-27 17:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:32.550198084 +0000 UTC m=+1747.822655436" watchObservedRunningTime="2025-11-27 17:38:32.566365535 +0000 UTC m=+1747.838822897" Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.623672 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7fd4-account-create-update-pg2kx"] Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.638826 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-vtslx"] Nov 27 17:38:32 crc kubenswrapper[4809]: I1127 17:38:32.735085 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-19b4-account-create-update-n9t2g"] Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.149952 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-h8g6t-config-rhl6b"] Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.168398 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-h8g6t-config-rhl6b"] Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.251080 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-h8g6t-config-9zqdq"] Nov 27 17:38:33 crc kubenswrapper[4809]: E1127 17:38:33.251654 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae0eae0-702d-485d-8438-db6f005ad77f" containerName="ovn-config" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.251718 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae0eae0-702d-485d-8438-db6f005ad77f" containerName="ovn-config" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.252109 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae0eae0-702d-485d-8438-db6f005ad77f" containerName="ovn-config" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.252880 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.262326 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.277643 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h8g6t-config-9zqdq"] Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.395876 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58gnb\" (UniqueName: \"kubernetes.io/projected/b41e6e14-c07d-4b95-94c9-49a0bdade734-kube-api-access-58gnb\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.396224 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-scripts\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.396266 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.396332 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run-ovn\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.396427 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-log-ovn\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.396475 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-additional-scripts\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.468849 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ae0eae0-702d-485d-8438-db6f005ad77f" path="/var/lib/kubelet/pods/1ae0eae0-702d-485d-8438-db6f005ad77f/volumes" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.498208 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58gnb\" (UniqueName: \"kubernetes.io/projected/b41e6e14-c07d-4b95-94c9-49a0bdade734-kube-api-access-58gnb\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.498277 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-scripts\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.498316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.498381 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run-ovn\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.498467 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-log-ovn\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.498511 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-additional-scripts\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.498748 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.498734 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-log-ovn\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.498850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run-ovn\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.499614 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-additional-scripts\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.500915 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-scripts\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.519502 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58gnb\" (UniqueName: \"kubernetes.io/projected/b41e6e14-c07d-4b95-94c9-49a0bdade734-kube-api-access-58gnb\") pod \"ovn-controller-h8g6t-config-9zqdq\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.527190 4809 generic.go:334] "Generic (PLEG): container finished" podID="4ae75b87-925b-4b0c-9a59-1d59f534f9f2" containerID="2c83e01330814afae94a2e95302d070a959f576dc9b7f275f32f672d4d8dd57f" exitCode=0 Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.527261 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vg6fv" event={"ID":"4ae75b87-925b-4b0c-9a59-1d59f534f9f2","Type":"ContainerDied","Data":"2c83e01330814afae94a2e95302d070a959f576dc9b7f275f32f672d4d8dd57f"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.530526 4809 generic.go:334] "Generic (PLEG): container finished" podID="b374716e-7e54-4be1-96bf-3c9e9bd379b8" containerID="a2ef4cd023c1bc6fc21cd49a10d176140642e6687e16293b03942033304efb52" exitCode=0 Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.530564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8bnvg" event={"ID":"b374716e-7e54-4be1-96bf-3c9e9bd379b8","Type":"ContainerDied","Data":"a2ef4cd023c1bc6fc21cd49a10d176140642e6687e16293b03942033304efb52"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.534526 4809 generic.go:334] "Generic (PLEG): container finished" podID="025088fd-c5dc-452e-aceb-c4a04cc09033" containerID="09d8633d799bef52f0183e342dc16ab5245ee464bc71186d4d6914dddc30cbc2" exitCode=0 Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.534570 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7fd4-account-create-update-pg2kx" event={"ID":"025088fd-c5dc-452e-aceb-c4a04cc09033","Type":"ContainerDied","Data":"09d8633d799bef52f0183e342dc16ab5245ee464bc71186d4d6914dddc30cbc2"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.534588 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7fd4-account-create-update-pg2kx" event={"ID":"025088fd-c5dc-452e-aceb-c4a04cc09033","Type":"ContainerStarted","Data":"b6835c167abfbb9d10d055a0fd3aa7793cefad9ec24e9716f9691c84d8cd928c"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.538901 4809 generic.go:334] "Generic (PLEG): container finished" podID="4dcd2971-5635-42ab-af5c-ee867a78bfa1" containerID="f65f913622348400865367354a71250bcd9fe05e5fe48157f3efa94651a00405" exitCode=0 Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.538974 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-vtslx" event={"ID":"4dcd2971-5635-42ab-af5c-ee867a78bfa1","Type":"ContainerDied","Data":"f65f913622348400865367354a71250bcd9fe05e5fe48157f3efa94651a00405"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.539001 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-vtslx" event={"ID":"4dcd2971-5635-42ab-af5c-ee867a78bfa1","Type":"ContainerStarted","Data":"826991dccf26fbd8640ce38ddede688d08593fdcd59a3706b5f7633caa24a891"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.541313 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5tlhg" event={"ID":"4891ffad-a908-49ac-97f5-196ae7a2e919","Type":"ContainerStarted","Data":"206d6e62d0fbe750886993dfbccf2b3c22b58693765e82e1c2ca56ba3123559a"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.548342 4809 generic.go:334] "Generic (PLEG): container finished" podID="c4a9f9d1-a45a-49a1-95bd-1681a1b220ff" containerID="aa81b99ab71174dca079f8cce2d31038d0c85915e7e7aa876ddb7b5d15cddbab" exitCode=0 Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.548430 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-87f3-account-create-update-klgkz" event={"ID":"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff","Type":"ContainerDied","Data":"aa81b99ab71174dca079f8cce2d31038d0c85915e7e7aa876ddb7b5d15cddbab"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.554209 4809 generic.go:334] "Generic (PLEG): container finished" podID="d44f63e5-7e83-4cd8-a488-0e88f7635938" containerID="800ddd6ba7a20a24f4be54afcd2a5622e7e8ccf7e7b8538fc7fc014b6e666719" exitCode=0 Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.554288 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-19b4-account-create-update-n9t2g" event={"ID":"d44f63e5-7e83-4cd8-a488-0e88f7635938","Type":"ContainerDied","Data":"800ddd6ba7a20a24f4be54afcd2a5622e7e8ccf7e7b8538fc7fc014b6e666719"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.554314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-19b4-account-create-update-n9t2g" event={"ID":"d44f63e5-7e83-4cd8-a488-0e88f7635938","Type":"ContainerStarted","Data":"f7fef5131df9ffbf567b3bb28a68480e4444f5ccaa3d0c8e59d961d6aab676f0"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.566219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" event={"ID":"d542ca5d-c018-4f64-96f0-e515d148c51c","Type":"ContainerStarted","Data":"76cdd523a6e9bc63ffd51c4fb46910fc4a38ea7c8f838aca7f6f6ccf855f3452"} Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.566396 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.575419 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:33 crc kubenswrapper[4809]: I1127 17:38:33.638504 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" podStartSLOduration=3.638482586 podStartE2EDuration="3.638482586s" podCreationTimestamp="2025-11-27 17:38:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:33.634566659 +0000 UTC m=+1748.907024011" watchObservedRunningTime="2025-11-27 17:38:33.638482586 +0000 UTC m=+1748.910939938" Nov 27 17:38:34 crc kubenswrapper[4809]: I1127 17:38:34.049967 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h8g6t-config-9zqdq"] Nov 27 17:38:34 crc kubenswrapper[4809]: W1127 17:38:34.059961 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb41e6e14_c07d_4b95_94c9_49a0bdade734.slice/crio-5d50846fa559f586b39038c35c33916bb71197c96baef9a21a2ad0a2581e3e86 WatchSource:0}: Error finding container 5d50846fa559f586b39038c35c33916bb71197c96baef9a21a2ad0a2581e3e86: Status 404 returned error can't find the container with id 5d50846fa559f586b39038c35c33916bb71197c96baef9a21a2ad0a2581e3e86 Nov 27 17:38:34 crc kubenswrapper[4809]: I1127 17:38:34.576580 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t-config-9zqdq" event={"ID":"b41e6e14-c07d-4b95-94c9-49a0bdade734","Type":"ContainerStarted","Data":"c15b4e7c3ade0d28dabdfa9a5c9de9de7e4eba3b3ffca27b1e5251c7acbbb3c7"} Nov 27 17:38:34 crc kubenswrapper[4809]: I1127 17:38:34.576959 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t-config-9zqdq" event={"ID":"b41e6e14-c07d-4b95-94c9-49a0bdade734","Type":"ContainerStarted","Data":"5d50846fa559f586b39038c35c33916bb71197c96baef9a21a2ad0a2581e3e86"} Nov 27 17:38:34 crc kubenswrapper[4809]: I1127 17:38:34.600065 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-h8g6t-config-9zqdq" podStartSLOduration=1.6000464380000001 podStartE2EDuration="1.600046438s" podCreationTimestamp="2025-11-27 17:38:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:34.597099228 +0000 UTC m=+1749.869556580" watchObservedRunningTime="2025-11-27 17:38:34.600046438 +0000 UTC m=+1749.872503800" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.155035 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.234964 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-h8g6t" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.247371 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-operator-scripts\") pod \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\" (UID: \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.247535 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wghhn\" (UniqueName: \"kubernetes.io/projected/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-kube-api-access-wghhn\") pod \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\" (UID: \"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.249896 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c4a9f9d1-a45a-49a1-95bd-1681a1b220ff" (UID: "c4a9f9d1-a45a-49a1-95bd-1681a1b220ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.266157 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-kube-api-access-wghhn" (OuterVolumeSpecName: "kube-api-access-wghhn") pod "c4a9f9d1-a45a-49a1-95bd-1681a1b220ff" (UID: "c4a9f9d1-a45a-49a1-95bd-1681a1b220ff"). InnerVolumeSpecName "kube-api-access-wghhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.350646 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wghhn\" (UniqueName: \"kubernetes.io/projected/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-kube-api-access-wghhn\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.350962 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.432320 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.438794 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.447150 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.471457 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.476145 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.558188 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr6vb\" (UniqueName: \"kubernetes.io/projected/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-kube-api-access-cr6vb\") pod \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\" (UID: \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.558684 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/025088fd-c5dc-452e-aceb-c4a04cc09033-operator-scripts\") pod \"025088fd-c5dc-452e-aceb-c4a04cc09033\" (UID: \"025088fd-c5dc-452e-aceb-c4a04cc09033\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.559260 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/025088fd-c5dc-452e-aceb-c4a04cc09033-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "025088fd-c5dc-452e-aceb-c4a04cc09033" (UID: "025088fd-c5dc-452e-aceb-c4a04cc09033"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.559461 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b7qp\" (UniqueName: \"kubernetes.io/projected/b374716e-7e54-4be1-96bf-3c9e9bd379b8-kube-api-access-4b7qp\") pod \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\" (UID: \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.559926 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b374716e-7e54-4be1-96bf-3c9e9bd379b8-operator-scripts\") pod \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\" (UID: \"b374716e-7e54-4be1-96bf-3c9e9bd379b8\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.560800 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktbnr\" (UniqueName: \"kubernetes.io/projected/4dcd2971-5635-42ab-af5c-ee867a78bfa1-kube-api-access-ktbnr\") pod \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\" (UID: \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.560949 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm5tg\" (UniqueName: \"kubernetes.io/projected/025088fd-c5dc-452e-aceb-c4a04cc09033-kube-api-access-jm5tg\") pod \"025088fd-c5dc-452e-aceb-c4a04cc09033\" (UID: \"025088fd-c5dc-452e-aceb-c4a04cc09033\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.560809 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b374716e-7e54-4be1-96bf-3c9e9bd379b8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b374716e-7e54-4be1-96bf-3c9e9bd379b8" (UID: "b374716e-7e54-4be1-96bf-3c9e9bd379b8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.561031 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d44f63e5-7e83-4cd8-a488-0e88f7635938-operator-scripts\") pod \"d44f63e5-7e83-4cd8-a488-0e88f7635938\" (UID: \"d44f63e5-7e83-4cd8-a488-0e88f7635938\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.561106 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lndj\" (UniqueName: \"kubernetes.io/projected/d44f63e5-7e83-4cd8-a488-0e88f7635938-kube-api-access-4lndj\") pod \"d44f63e5-7e83-4cd8-a488-0e88f7635938\" (UID: \"d44f63e5-7e83-4cd8-a488-0e88f7635938\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.561180 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-operator-scripts\") pod \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\" (UID: \"4ae75b87-925b-4b0c-9a59-1d59f534f9f2\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.561212 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcd2971-5635-42ab-af5c-ee867a78bfa1-operator-scripts\") pod \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\" (UID: \"4dcd2971-5635-42ab-af5c-ee867a78bfa1\") " Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.561551 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ae75b87-925b-4b0c-9a59-1d59f534f9f2" (UID: "4ae75b87-925b-4b0c-9a59-1d59f534f9f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.561926 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d44f63e5-7e83-4cd8-a488-0e88f7635938-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d44f63e5-7e83-4cd8-a488-0e88f7635938" (UID: "d44f63e5-7e83-4cd8-a488-0e88f7635938"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.561955 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dcd2971-5635-42ab-af5c-ee867a78bfa1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4dcd2971-5635-42ab-af5c-ee867a78bfa1" (UID: "4dcd2971-5635-42ab-af5c-ee867a78bfa1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.562101 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/025088fd-c5dc-452e-aceb-c4a04cc09033-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.563576 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b374716e-7e54-4be1-96bf-3c9e9bd379b8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.563646 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.564236 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-kube-api-access-cr6vb" (OuterVolumeSpecName: "kube-api-access-cr6vb") pod "4ae75b87-925b-4b0c-9a59-1d59f534f9f2" (UID: "4ae75b87-925b-4b0c-9a59-1d59f534f9f2"). InnerVolumeSpecName "kube-api-access-cr6vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.564610 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b374716e-7e54-4be1-96bf-3c9e9bd379b8-kube-api-access-4b7qp" (OuterVolumeSpecName: "kube-api-access-4b7qp") pod "b374716e-7e54-4be1-96bf-3c9e9bd379b8" (UID: "b374716e-7e54-4be1-96bf-3c9e9bd379b8"). InnerVolumeSpecName "kube-api-access-4b7qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.566078 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dcd2971-5635-42ab-af5c-ee867a78bfa1-kube-api-access-ktbnr" (OuterVolumeSpecName: "kube-api-access-ktbnr") pod "4dcd2971-5635-42ab-af5c-ee867a78bfa1" (UID: "4dcd2971-5635-42ab-af5c-ee867a78bfa1"). InnerVolumeSpecName "kube-api-access-ktbnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.567041 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/025088fd-c5dc-452e-aceb-c4a04cc09033-kube-api-access-jm5tg" (OuterVolumeSpecName: "kube-api-access-jm5tg") pod "025088fd-c5dc-452e-aceb-c4a04cc09033" (UID: "025088fd-c5dc-452e-aceb-c4a04cc09033"). InnerVolumeSpecName "kube-api-access-jm5tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.568532 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d44f63e5-7e83-4cd8-a488-0e88f7635938-kube-api-access-4lndj" (OuterVolumeSpecName: "kube-api-access-4lndj") pod "d44f63e5-7e83-4cd8-a488-0e88f7635938" (UID: "d44f63e5-7e83-4cd8-a488-0e88f7635938"). InnerVolumeSpecName "kube-api-access-4lndj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.587294 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8bnvg" event={"ID":"b374716e-7e54-4be1-96bf-3c9e9bd379b8","Type":"ContainerDied","Data":"664dcadb1a5c8988f2e6c2758480e02cbb4f97208609859b013968d5685b13d4"} Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.587338 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="664dcadb1a5c8988f2e6c2758480e02cbb4f97208609859b013968d5685b13d4" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.587413 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8bnvg" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.594435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7fd4-account-create-update-pg2kx" event={"ID":"025088fd-c5dc-452e-aceb-c4a04cc09033","Type":"ContainerDied","Data":"b6835c167abfbb9d10d055a0fd3aa7793cefad9ec24e9716f9691c84d8cd928c"} Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.594483 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6835c167abfbb9d10d055a0fd3aa7793cefad9ec24e9716f9691c84d8cd928c" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.594458 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7fd4-account-create-update-pg2kx" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.596111 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-vtslx" event={"ID":"4dcd2971-5635-42ab-af5c-ee867a78bfa1","Type":"ContainerDied","Data":"826991dccf26fbd8640ce38ddede688d08593fdcd59a3706b5f7633caa24a891"} Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.596151 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="826991dccf26fbd8640ce38ddede688d08593fdcd59a3706b5f7633caa24a891" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.596266 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-vtslx" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.597531 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-87f3-account-create-update-klgkz" event={"ID":"c4a9f9d1-a45a-49a1-95bd-1681a1b220ff","Type":"ContainerDied","Data":"9020cf03049accbea57df8cee91fa0bfad6372f78d713675a1fe35b861897263"} Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.597561 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9020cf03049accbea57df8cee91fa0bfad6372f78d713675a1fe35b861897263" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.597538 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-87f3-account-create-update-klgkz" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.599093 4809 generic.go:334] "Generic (PLEG): container finished" podID="b41e6e14-c07d-4b95-94c9-49a0bdade734" containerID="c15b4e7c3ade0d28dabdfa9a5c9de9de7e4eba3b3ffca27b1e5251c7acbbb3c7" exitCode=0 Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.599173 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t-config-9zqdq" event={"ID":"b41e6e14-c07d-4b95-94c9-49a0bdade734","Type":"ContainerDied","Data":"c15b4e7c3ade0d28dabdfa9a5c9de9de7e4eba3b3ffca27b1e5251c7acbbb3c7"} Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.610935 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-19b4-account-create-update-n9t2g" event={"ID":"d44f63e5-7e83-4cd8-a488-0e88f7635938","Type":"ContainerDied","Data":"f7fef5131df9ffbf567b3bb28a68480e4444f5ccaa3d0c8e59d961d6aab676f0"} Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.610976 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7fef5131df9ffbf567b3bb28a68480e4444f5ccaa3d0c8e59d961d6aab676f0" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.611031 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-19b4-account-create-update-n9t2g" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.621229 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vg6fv" event={"ID":"4ae75b87-925b-4b0c-9a59-1d59f534f9f2","Type":"ContainerDied","Data":"b5db5fa7585cfeddac410e783daadc37f302e1e8d585a10d2d803bf4776820c0"} Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.621647 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5db5fa7585cfeddac410e783daadc37f302e1e8d585a10d2d803bf4776820c0" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.621878 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vg6fv" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.666258 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b7qp\" (UniqueName: \"kubernetes.io/projected/b374716e-7e54-4be1-96bf-3c9e9bd379b8-kube-api-access-4b7qp\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.666846 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktbnr\" (UniqueName: \"kubernetes.io/projected/4dcd2971-5635-42ab-af5c-ee867a78bfa1-kube-api-access-ktbnr\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.666864 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm5tg\" (UniqueName: \"kubernetes.io/projected/025088fd-c5dc-452e-aceb-c4a04cc09033-kube-api-access-jm5tg\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.666876 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d44f63e5-7e83-4cd8-a488-0e88f7635938-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.666888 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lndj\" (UniqueName: \"kubernetes.io/projected/d44f63e5-7e83-4cd8-a488-0e88f7635938-kube-api-access-4lndj\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.666900 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcd2971-5635-42ab-af5c-ee867a78bfa1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:35 crc kubenswrapper[4809]: I1127 17:38:35.666913 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr6vb\" (UniqueName: \"kubernetes.io/projected/4ae75b87-925b-4b0c-9a59-1d59f534f9f2-kube-api-access-cr6vb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:36 crc kubenswrapper[4809]: I1127 17:38:36.994338 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.089560 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-log-ovn\") pod \"b41e6e14-c07d-4b95-94c9-49a0bdade734\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.089703 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run-ovn\") pod \"b41e6e14-c07d-4b95-94c9-49a0bdade734\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.089703 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b41e6e14-c07d-4b95-94c9-49a0bdade734" (UID: "b41e6e14-c07d-4b95-94c9-49a0bdade734"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.089799 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b41e6e14-c07d-4b95-94c9-49a0bdade734" (UID: "b41e6e14-c07d-4b95-94c9-49a0bdade734"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.089921 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-additional-scripts\") pod \"b41e6e14-c07d-4b95-94c9-49a0bdade734\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.090652 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b41e6e14-c07d-4b95-94c9-49a0bdade734" (UID: "b41e6e14-c07d-4b95-94c9-49a0bdade734"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.090719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-scripts\") pod \"b41e6e14-c07d-4b95-94c9-49a0bdade734\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.091382 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-scripts" (OuterVolumeSpecName: "scripts") pod "b41e6e14-c07d-4b95-94c9-49a0bdade734" (UID: "b41e6e14-c07d-4b95-94c9-49a0bdade734"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.091598 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run\") pod \"b41e6e14-c07d-4b95-94c9-49a0bdade734\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.091663 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run" (OuterVolumeSpecName: "var-run") pod "b41e6e14-c07d-4b95-94c9-49a0bdade734" (UID: "b41e6e14-c07d-4b95-94c9-49a0bdade734"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.091687 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58gnb\" (UniqueName: \"kubernetes.io/projected/b41e6e14-c07d-4b95-94c9-49a0bdade734-kube-api-access-58gnb\") pod \"b41e6e14-c07d-4b95-94c9-49a0bdade734\" (UID: \"b41e6e14-c07d-4b95-94c9-49a0bdade734\") " Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.092198 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.092220 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.092231 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b41e6e14-c07d-4b95-94c9-49a0bdade734-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.092242 4809 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.092255 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b41e6e14-c07d-4b95-94c9-49a0bdade734-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.096263 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b41e6e14-c07d-4b95-94c9-49a0bdade734-kube-api-access-58gnb" (OuterVolumeSpecName: "kube-api-access-58gnb") pod "b41e6e14-c07d-4b95-94c9-49a0bdade734" (UID: "b41e6e14-c07d-4b95-94c9-49a0bdade734"). InnerVolumeSpecName "kube-api-access-58gnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.193957 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58gnb\" (UniqueName: \"kubernetes.io/projected/b41e6e14-c07d-4b95-94c9-49a0bdade734-kube-api-access-58gnb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.646009 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h8g6t-config-9zqdq" event={"ID":"b41e6e14-c07d-4b95-94c9-49a0bdade734","Type":"ContainerDied","Data":"5d50846fa559f586b39038c35c33916bb71197c96baef9a21a2ad0a2581e3e86"} Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.646378 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d50846fa559f586b39038c35c33916bb71197c96baef9a21a2ad0a2581e3e86" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.646204 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h8g6t-config-9zqdq" Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.666398 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-h8g6t-config-9zqdq"] Nov 27 17:38:37 crc kubenswrapper[4809]: I1127 17:38:37.677435 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-h8g6t-config-9zqdq"] Nov 27 17:38:38 crc kubenswrapper[4809]: I1127 17:38:38.457696 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:38:38 crc kubenswrapper[4809]: E1127 17:38:38.458185 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:38:39 crc kubenswrapper[4809]: I1127 17:38:39.469343 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b41e6e14-c07d-4b95-94c9-49a0bdade734" path="/var/lib/kubelet/pods/b41e6e14-c07d-4b95-94c9-49a0bdade734/volumes" Nov 27 17:38:40 crc kubenswrapper[4809]: I1127 17:38:40.670033 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5tlhg" event={"ID":"4891ffad-a908-49ac-97f5-196ae7a2e919","Type":"ContainerStarted","Data":"186036199ff8fe73257a5d42a2cbabb177d45c47333b1d8d3c20053c2fd4e047"} Nov 27 17:38:40 crc kubenswrapper[4809]: I1127 17:38:40.687345 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-5tlhg" podStartSLOduration=2.364698335 podStartE2EDuration="9.687328402s" podCreationTimestamp="2025-11-27 17:38:31 +0000 UTC" firstStartedPulling="2025-11-27 17:38:32.516908179 +0000 UTC m=+1747.789365531" lastFinishedPulling="2025-11-27 17:38:39.839538096 +0000 UTC m=+1755.111995598" observedRunningTime="2025-11-27 17:38:40.682763717 +0000 UTC m=+1755.955221079" watchObservedRunningTime="2025-11-27 17:38:40.687328402 +0000 UTC m=+1755.959785754" Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.263887 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.318652 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-wvhh5"] Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.319139 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-wvhh5" podUID="a4cefed0-40bf-451b-bc22-609d313f1d67" containerName="dnsmasq-dns" containerID="cri-o://86cd05b94e76ae5473fc96506b128cf3eb2af7f27861053ba3d0d7e6e2e362e1" gracePeriod=10 Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.681577 4809 generic.go:334] "Generic (PLEG): container finished" podID="a4cefed0-40bf-451b-bc22-609d313f1d67" containerID="86cd05b94e76ae5473fc96506b128cf3eb2af7f27861053ba3d0d7e6e2e362e1" exitCode=0 Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.681649 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-wvhh5" event={"ID":"a4cefed0-40bf-451b-bc22-609d313f1d67","Type":"ContainerDied","Data":"86cd05b94e76ae5473fc96506b128cf3eb2af7f27861053ba3d0d7e6e2e362e1"} Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.683754 4809 generic.go:334] "Generic (PLEG): container finished" podID="acc5484a-4058-4a5d-97a0-e84aaa75905f" containerID="b52055a16d85e6af2503992f7364d1eb40fb9f67870308ae59d2dc63673b2fd0" exitCode=0 Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.684617 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bm4mw" event={"ID":"acc5484a-4058-4a5d-97a0-e84aaa75905f","Type":"ContainerDied","Data":"b52055a16d85e6af2503992f7364d1eb40fb9f67870308ae59d2dc63673b2fd0"} Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.850304 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.984685 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-sb\") pod \"a4cefed0-40bf-451b-bc22-609d313f1d67\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.984792 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-config\") pod \"a4cefed0-40bf-451b-bc22-609d313f1d67\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.984840 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-dns-svc\") pod \"a4cefed0-40bf-451b-bc22-609d313f1d67\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.984943 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-nb\") pod \"a4cefed0-40bf-451b-bc22-609d313f1d67\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.985031 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt97w\" (UniqueName: \"kubernetes.io/projected/a4cefed0-40bf-451b-bc22-609d313f1d67-kube-api-access-qt97w\") pod \"a4cefed0-40bf-451b-bc22-609d313f1d67\" (UID: \"a4cefed0-40bf-451b-bc22-609d313f1d67\") " Nov 27 17:38:41 crc kubenswrapper[4809]: I1127 17:38:41.991279 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4cefed0-40bf-451b-bc22-609d313f1d67-kube-api-access-qt97w" (OuterVolumeSpecName: "kube-api-access-qt97w") pod "a4cefed0-40bf-451b-bc22-609d313f1d67" (UID: "a4cefed0-40bf-451b-bc22-609d313f1d67"). InnerVolumeSpecName "kube-api-access-qt97w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.032712 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-config" (OuterVolumeSpecName: "config") pod "a4cefed0-40bf-451b-bc22-609d313f1d67" (UID: "a4cefed0-40bf-451b-bc22-609d313f1d67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.033892 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a4cefed0-40bf-451b-bc22-609d313f1d67" (UID: "a4cefed0-40bf-451b-bc22-609d313f1d67"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.038216 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a4cefed0-40bf-451b-bc22-609d313f1d67" (UID: "a4cefed0-40bf-451b-bc22-609d313f1d67"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.042894 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a4cefed0-40bf-451b-bc22-609d313f1d67" (UID: "a4cefed0-40bf-451b-bc22-609d313f1d67"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.086521 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt97w\" (UniqueName: \"kubernetes.io/projected/a4cefed0-40bf-451b-bc22-609d313f1d67-kube-api-access-qt97w\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.086555 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.086566 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.086574 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.086581 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4cefed0-40bf-451b-bc22-609d313f1d67-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.694291 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-wvhh5" event={"ID":"a4cefed0-40bf-451b-bc22-609d313f1d67","Type":"ContainerDied","Data":"818b5edf05f854e47b5feaf828f85cbddcd8aa65840cd1cb16f4dd3e2dfb0e16"} Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.694359 4809 scope.go:117] "RemoveContainer" containerID="86cd05b94e76ae5473fc96506b128cf3eb2af7f27861053ba3d0d7e6e2e362e1" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.694466 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-wvhh5" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.732268 4809 scope.go:117] "RemoveContainer" containerID="8bcd7f10799eb9e4a1bc14f47ab770911f873ea6a7a03f47e203b16cf4c55867" Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.734597 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-wvhh5"] Nov 27 17:38:42 crc kubenswrapper[4809]: I1127 17:38:42.744605 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-wvhh5"] Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.403911 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.474521 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4cefed0-40bf-451b-bc22-609d313f1d67" path="/var/lib/kubelet/pods/a4cefed0-40bf-451b-bc22-609d313f1d67/volumes" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.514872 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-combined-ca-bundle\") pod \"acc5484a-4058-4a5d-97a0-e84aaa75905f\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.515049 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fcwg\" (UniqueName: \"kubernetes.io/projected/acc5484a-4058-4a5d-97a0-e84aaa75905f-kube-api-access-9fcwg\") pod \"acc5484a-4058-4a5d-97a0-e84aaa75905f\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.515148 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-config-data\") pod \"acc5484a-4058-4a5d-97a0-e84aaa75905f\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.515250 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-db-sync-config-data\") pod \"acc5484a-4058-4a5d-97a0-e84aaa75905f\" (UID: \"acc5484a-4058-4a5d-97a0-e84aaa75905f\") " Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.521039 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "acc5484a-4058-4a5d-97a0-e84aaa75905f" (UID: "acc5484a-4058-4a5d-97a0-e84aaa75905f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.521675 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acc5484a-4058-4a5d-97a0-e84aaa75905f-kube-api-access-9fcwg" (OuterVolumeSpecName: "kube-api-access-9fcwg") pod "acc5484a-4058-4a5d-97a0-e84aaa75905f" (UID: "acc5484a-4058-4a5d-97a0-e84aaa75905f"). InnerVolumeSpecName "kube-api-access-9fcwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.545504 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "acc5484a-4058-4a5d-97a0-e84aaa75905f" (UID: "acc5484a-4058-4a5d-97a0-e84aaa75905f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.573210 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-config-data" (OuterVolumeSpecName: "config-data") pod "acc5484a-4058-4a5d-97a0-e84aaa75905f" (UID: "acc5484a-4058-4a5d-97a0-e84aaa75905f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.618160 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fcwg\" (UniqueName: \"kubernetes.io/projected/acc5484a-4058-4a5d-97a0-e84aaa75905f-kube-api-access-9fcwg\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.618219 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.618235 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.618251 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc5484a-4058-4a5d-97a0-e84aaa75905f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.706975 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bm4mw" event={"ID":"acc5484a-4058-4a5d-97a0-e84aaa75905f","Type":"ContainerDied","Data":"646c934206be17400d57ed844c6cddb68cdf6fe2a393cf3143772c283434eeb1"} Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.707029 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="646c934206be17400d57ed844c6cddb68cdf6fe2a393cf3143772c283434eeb1" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.706999 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bm4mw" Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.709266 4809 generic.go:334] "Generic (PLEG): container finished" podID="4891ffad-a908-49ac-97f5-196ae7a2e919" containerID="186036199ff8fe73257a5d42a2cbabb177d45c47333b1d8d3c20053c2fd4e047" exitCode=0 Nov 27 17:38:43 crc kubenswrapper[4809]: I1127 17:38:43.709296 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5tlhg" event={"ID":"4891ffad-a908-49ac-97f5-196ae7a2e919","Type":"ContainerDied","Data":"186036199ff8fe73257a5d42a2cbabb177d45c47333b1d8d3c20053c2fd4e047"} Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.085170 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jhdt8"] Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.087557 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4a9f9d1-a45a-49a1-95bd-1681a1b220ff" containerName="mariadb-account-create-update" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.087759 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4a9f9d1-a45a-49a1-95bd-1681a1b220ff" containerName="mariadb-account-create-update" Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.087849 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acc5484a-4058-4a5d-97a0-e84aaa75905f" containerName="glance-db-sync" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.087916 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="acc5484a-4058-4a5d-97a0-e84aaa75905f" containerName="glance-db-sync" Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.087997 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4cefed0-40bf-451b-bc22-609d313f1d67" containerName="init" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.088070 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4cefed0-40bf-451b-bc22-609d313f1d67" containerName="init" Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.088141 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae75b87-925b-4b0c-9a59-1d59f534f9f2" containerName="mariadb-database-create" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.088244 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae75b87-925b-4b0c-9a59-1d59f534f9f2" containerName="mariadb-database-create" Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.088322 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025088fd-c5dc-452e-aceb-c4a04cc09033" containerName="mariadb-account-create-update" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.088396 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="025088fd-c5dc-452e-aceb-c4a04cc09033" containerName="mariadb-account-create-update" Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.088475 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dcd2971-5635-42ab-af5c-ee867a78bfa1" containerName="mariadb-database-create" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.088541 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dcd2971-5635-42ab-af5c-ee867a78bfa1" containerName="mariadb-database-create" Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.088603 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44f63e5-7e83-4cd8-a488-0e88f7635938" containerName="mariadb-account-create-update" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.088662 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44f63e5-7e83-4cd8-a488-0e88f7635938" containerName="mariadb-account-create-update" Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.088731 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b41e6e14-c07d-4b95-94c9-49a0bdade734" containerName="ovn-config" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.088843 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b41e6e14-c07d-4b95-94c9-49a0bdade734" containerName="ovn-config" Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.088910 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b374716e-7e54-4be1-96bf-3c9e9bd379b8" containerName="mariadb-database-create" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.088974 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b374716e-7e54-4be1-96bf-3c9e9bd379b8" containerName="mariadb-database-create" Nov 27 17:38:44 crc kubenswrapper[4809]: E1127 17:38:44.089054 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4cefed0-40bf-451b-bc22-609d313f1d67" containerName="dnsmasq-dns" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.089855 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4cefed0-40bf-451b-bc22-609d313f1d67" containerName="dnsmasq-dns" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.090166 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4a9f9d1-a45a-49a1-95bd-1681a1b220ff" containerName="mariadb-account-create-update" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.090254 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b374716e-7e54-4be1-96bf-3c9e9bd379b8" containerName="mariadb-database-create" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.090332 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="025088fd-c5dc-452e-aceb-c4a04cc09033" containerName="mariadb-account-create-update" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.090433 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ae75b87-925b-4b0c-9a59-1d59f534f9f2" containerName="mariadb-database-create" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.090529 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4cefed0-40bf-451b-bc22-609d313f1d67" containerName="dnsmasq-dns" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.090609 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d44f63e5-7e83-4cd8-a488-0e88f7635938" containerName="mariadb-account-create-update" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.090697 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b41e6e14-c07d-4b95-94c9-49a0bdade734" containerName="ovn-config" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.090797 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dcd2971-5635-42ab-af5c-ee867a78bfa1" containerName="mariadb-database-create" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.090907 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="acc5484a-4058-4a5d-97a0-e84aaa75905f" containerName="glance-db-sync" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.092359 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.097642 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jhdt8"] Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.243213 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j24gm\" (UniqueName: \"kubernetes.io/projected/ab62f250-6ba2-4c46-bda1-061a52f9928a-kube-api-access-j24gm\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.243295 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.243341 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-config\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.243373 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.243399 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.243446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.344546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.344980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j24gm\" (UniqueName: \"kubernetes.io/projected/ab62f250-6ba2-4c46-bda1-061a52f9928a-kube-api-access-j24gm\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.345021 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.345060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-config\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.345090 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.345115 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.345454 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.346023 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.346037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.346206 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-config\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.346754 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.363585 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j24gm\" (UniqueName: \"kubernetes.io/projected/ab62f250-6ba2-4c46-bda1-061a52f9928a-kube-api-access-j24gm\") pod \"dnsmasq-dns-7ff5475cc9-jhdt8\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.417003 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:44 crc kubenswrapper[4809]: I1127 17:38:44.776827 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jhdt8"] Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.200384 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.365527 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-combined-ca-bundle\") pod \"4891ffad-a908-49ac-97f5-196ae7a2e919\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.365666 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-config-data\") pod \"4891ffad-a908-49ac-97f5-196ae7a2e919\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.365957 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjgd6\" (UniqueName: \"kubernetes.io/projected/4891ffad-a908-49ac-97f5-196ae7a2e919-kube-api-access-wjgd6\") pod \"4891ffad-a908-49ac-97f5-196ae7a2e919\" (UID: \"4891ffad-a908-49ac-97f5-196ae7a2e919\") " Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.370246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4891ffad-a908-49ac-97f5-196ae7a2e919-kube-api-access-wjgd6" (OuterVolumeSpecName: "kube-api-access-wjgd6") pod "4891ffad-a908-49ac-97f5-196ae7a2e919" (UID: "4891ffad-a908-49ac-97f5-196ae7a2e919"). InnerVolumeSpecName "kube-api-access-wjgd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.390385 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4891ffad-a908-49ac-97f5-196ae7a2e919" (UID: "4891ffad-a908-49ac-97f5-196ae7a2e919"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.421425 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-config-data" (OuterVolumeSpecName: "config-data") pod "4891ffad-a908-49ac-97f5-196ae7a2e919" (UID: "4891ffad-a908-49ac-97f5-196ae7a2e919"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.467367 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjgd6\" (UniqueName: \"kubernetes.io/projected/4891ffad-a908-49ac-97f5-196ae7a2e919-kube-api-access-wjgd6\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.467392 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.467402 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4891ffad-a908-49ac-97f5-196ae7a2e919-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.736472 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5tlhg" event={"ID":"4891ffad-a908-49ac-97f5-196ae7a2e919","Type":"ContainerDied","Data":"206d6e62d0fbe750886993dfbccf2b3c22b58693765e82e1c2ca56ba3123559a"} Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.736705 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="206d6e62d0fbe750886993dfbccf2b3c22b58693765e82e1c2ca56ba3123559a" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.736505 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5tlhg" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.737933 4809 generic.go:334] "Generic (PLEG): container finished" podID="ab62f250-6ba2-4c46-bda1-061a52f9928a" containerID="14d3407f4fca243a7afec7ae5ed60078840ad3ffbecfed5804c64ed4362eee63" exitCode=0 Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.737978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" event={"ID":"ab62f250-6ba2-4c46-bda1-061a52f9928a","Type":"ContainerDied","Data":"14d3407f4fca243a7afec7ae5ed60078840ad3ffbecfed5804c64ed4362eee63"} Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.738006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" event={"ID":"ab62f250-6ba2-4c46-bda1-061a52f9928a","Type":"ContainerStarted","Data":"6c54205c21810dfdcf8581df46a6d8215bfa7ecbe2bfc4d242f751abe958d54a"} Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.953337 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jhdt8"] Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.970903 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-lv8h9"] Nov 27 17:38:45 crc kubenswrapper[4809]: E1127 17:38:45.971289 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4891ffad-a908-49ac-97f5-196ae7a2e919" containerName="keystone-db-sync" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.971305 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4891ffad-a908-49ac-97f5-196ae7a2e919" containerName="keystone-db-sync" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.971467 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4891ffad-a908-49ac-97f5-196ae7a2e919" containerName="keystone-db-sync" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.972135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.977272 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.977477 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.977621 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.978271 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 17:38:45 crc kubenswrapper[4809]: I1127 17:38:45.978397 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-l4mwz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.005847 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lv8h9"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.037638 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.039464 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.081893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-scripts\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.081973 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-config-data\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.082072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89bz8\" (UniqueName: \"kubernetes.io/projected/ae4b6101-b917-4224-bc83-0b8e9327013c-kube-api-access-89bz8\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.082126 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-combined-ca-bundle\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.082149 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-credential-keys\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.082246 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-fernet-keys\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.094502 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9"] Nov 27 17:38:46 crc kubenswrapper[4809]: E1127 17:38:46.151857 4809 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 27 17:38:46 crc kubenswrapper[4809]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/ab62f250-6ba2-4c46-bda1-061a52f9928a/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 27 17:38:46 crc kubenswrapper[4809]: > podSandboxID="6c54205c21810dfdcf8581df46a6d8215bfa7ecbe2bfc4d242f751abe958d54a" Nov 27 17:38:46 crc kubenswrapper[4809]: E1127 17:38:46.152007 4809 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 27 17:38:46 crc kubenswrapper[4809]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64h68fh95h595h67fh597hfch57ch68fh5ffh6hf4h689h659h569h65bh67bh65dh594h64h5d6hd8h5bfh9fh5c4h676h5cdh56h8bh569h664h645q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j24gm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7ff5475cc9-jhdt8_openstack(ab62f250-6ba2-4c46-bda1-061a52f9928a): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/ab62f250-6ba2-4c46-bda1-061a52f9928a/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 27 17:38:46 crc kubenswrapper[4809]: > logger="UnhandledError" Nov 27 17:38:46 crc kubenswrapper[4809]: E1127 17:38:46.155407 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/ab62f250-6ba2-4c46-bda1-061a52f9928a/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" podUID="ab62f250-6ba2-4c46-bda1-061a52f9928a" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.202675 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-config-data\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.202811 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89bz8\" (UniqueName: \"kubernetes.io/projected/ae4b6101-b917-4224-bc83-0b8e9327013c-kube-api-access-89bz8\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.202900 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.202954 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-combined-ca-bundle\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.202987 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-credential-keys\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.203153 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.203176 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4746t\" (UniqueName: \"kubernetes.io/projected/b91aa6be-d5c9-4342-80b5-69011df32c42-kube-api-access-4746t\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.205968 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-t2gnx"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.206803 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.206893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-config\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.206916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.207082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-fernet-keys\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.207157 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-scripts\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.224693 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-fernet-keys\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.241428 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.247463 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89bz8\" (UniqueName: \"kubernetes.io/projected/ae4b6101-b917-4224-bc83-0b8e9327013c-kube-api-access-89bz8\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.256086 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2vvhs" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.257455 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.262216 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-scripts\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.266087 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.270944 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-config-data\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.277680 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-combined-ca-bundle\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.283081 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-t2gnx"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.292509 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-credential-keys\") pod \"keystone-bootstrap-lv8h9\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.311029 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-97thg"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.317602 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.345308 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.348757 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.348814 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.349464 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wngmx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.349583 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.349671 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.349708 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-config\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.350055 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.350161 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.350177 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4746t\" (UniqueName: \"kubernetes.io/projected/b91aa6be-d5c9-4342-80b5-69011df32c42-kube-api-access-4746t\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.350638 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.351145 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.351339 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.351719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.351873 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-config\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.352680 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-97thg"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.384527 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-rhmfz"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.386142 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.390609 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-lr45g" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.393240 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.393682 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.430500 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4746t\" (UniqueName: \"kubernetes.io/projected/b91aa6be-d5c9-4342-80b5-69011df32c42-kube-api-access-4746t\") pod \"dnsmasq-dns-5c5cc7c5ff-mwtf9\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.452219 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3b3964-4402-49a4-86ba-e6e7909082ec-etc-machine-id\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.452276 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-db-sync-config-data\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.452304 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwdz6\" (UniqueName: \"kubernetes.io/projected/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-kube-api-access-vwdz6\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.452319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-combined-ca-bundle\") pod \"neutron-db-sync-97thg\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.452337 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-logs\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.452494 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-scripts\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.454159 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-rhmfz"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.454226 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-config-data\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.454377 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-config-data\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.454504 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-combined-ca-bundle\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.454573 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrv44\" (UniqueName: \"kubernetes.io/projected/9a3b3964-4402-49a4-86ba-e6e7909082ec-kube-api-access-hrv44\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.454597 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-combined-ca-bundle\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.454649 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-scripts\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.454690 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kk75\" (UniqueName: \"kubernetes.io/projected/c184a0c3-5bb0-49f0-966f-fb565fd13202-kube-api-access-5kk75\") pod \"neutron-db-sync-97thg\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.454810 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-config\") pod \"neutron-db-sync-97thg\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.472199 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.472904 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.493020 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-n7gbr"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.494568 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.521138 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ktqz6"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.526921 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.538168 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qsvkg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.546557 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.556071 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ktqz6"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557049 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spbpw\" (UniqueName: \"kubernetes.io/projected/57b62f0e-7576-4e57-9993-30000f0eee5a-kube-api-access-spbpw\") pod \"barbican-db-sync-ktqz6\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557085 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-combined-ca-bundle\") pod \"barbican-db-sync-ktqz6\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557126 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-config-data\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557165 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-combined-ca-bundle\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557187 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-combined-ca-bundle\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrv44\" (UniqueName: \"kubernetes.io/projected/9a3b3964-4402-49a4-86ba-e6e7909082ec-kube-api-access-hrv44\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557222 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-scripts\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557290 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kk75\" (UniqueName: \"kubernetes.io/projected/c184a0c3-5bb0-49f0-966f-fb565fd13202-kube-api-access-5kk75\") pod \"neutron-db-sync-97thg\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557310 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-config\") pod \"neutron-db-sync-97thg\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557352 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-config\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557382 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3b3964-4402-49a4-86ba-e6e7909082ec-etc-machine-id\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557398 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-db-sync-config-data\") pod \"barbican-db-sync-ktqz6\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557444 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rf9s\" (UniqueName: \"kubernetes.io/projected/51a633d0-f7b6-4b1c-b609-7f9684069609-kube-api-access-5rf9s\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-db-sync-config-data\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557502 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwdz6\" (UniqueName: \"kubernetes.io/projected/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-kube-api-access-vwdz6\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-combined-ca-bundle\") pod \"neutron-db-sync-97thg\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557539 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557555 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-logs\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557570 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-scripts\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557606 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-config-data\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.557621 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.561128 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3b3964-4402-49a4-86ba-e6e7909082ec-etc-machine-id\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.561481 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-logs\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.576064 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-db-sync-config-data\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.576064 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-scripts\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.576430 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-scripts\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.576769 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-config-data\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.577898 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-config\") pod \"neutron-db-sync-97thg\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.591002 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-config-data\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.593631 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-combined-ca-bundle\") pod \"neutron-db-sync-97thg\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.600352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwdz6\" (UniqueName: \"kubernetes.io/projected/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-kube-api-access-vwdz6\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.600375 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-combined-ca-bundle\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.601128 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kk75\" (UniqueName: \"kubernetes.io/projected/c184a0c3-5bb0-49f0-966f-fb565fd13202-kube-api-access-5kk75\") pod \"neutron-db-sync-97thg\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.601607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrv44\" (UniqueName: \"kubernetes.io/projected/9a3b3964-4402-49a4-86ba-e6e7909082ec-kube-api-access-hrv44\") pod \"cinder-db-sync-t2gnx\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.603524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-combined-ca-bundle\") pod \"placement-db-sync-rhmfz\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.603959 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.604898 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-n7gbr"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.623485 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.631565 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.633854 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.639571 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.640125 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.659573 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.659624 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spbpw\" (UniqueName: \"kubernetes.io/projected/57b62f0e-7576-4e57-9993-30000f0eee5a-kube-api-access-spbpw\") pod \"barbican-db-sync-ktqz6\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.659647 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-combined-ca-bundle\") pod \"barbican-db-sync-ktqz6\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.659666 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.659707 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-run-httpd\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vlg4\" (UniqueName: \"kubernetes.io/projected/596ec651-2088-4459-b0dd-224c67150b5d-kube-api-access-2vlg4\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660465 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660485 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-scripts\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660514 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-config\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660541 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-log-httpd\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-db-sync-config-data\") pod \"barbican-db-sync-ktqz6\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660580 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rf9s\" (UniqueName: \"kubernetes.io/projected/51a633d0-f7b6-4b1c-b609-7f9684069609-kube-api-access-5rf9s\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660595 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-config-data\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660624 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.660641 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.661537 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.662852 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-config\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.663133 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.663235 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.663358 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.665676 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-combined-ca-bundle\") pod \"barbican-db-sync-ktqz6\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.680219 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-db-sync-config-data\") pod \"barbican-db-sync-ktqz6\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.690363 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-97thg" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.691381 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rf9s\" (UniqueName: \"kubernetes.io/projected/51a633d0-f7b6-4b1c-b609-7f9684069609-kube-api-access-5rf9s\") pod \"dnsmasq-dns-8b5c85b87-n7gbr\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.698304 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spbpw\" (UniqueName: \"kubernetes.io/projected/57b62f0e-7576-4e57-9993-30000f0eee5a-kube-api-access-spbpw\") pod \"barbican-db-sync-ktqz6\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.761777 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vlg4\" (UniqueName: \"kubernetes.io/projected/596ec651-2088-4459-b0dd-224c67150b5d-kube-api-access-2vlg4\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.761854 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.761874 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-scripts\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.761911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-log-httpd\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.761980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-config-data\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.762043 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.762095 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-run-httpd\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.762613 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-run-httpd\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.763423 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-log-httpd\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.768704 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.777204 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-rhmfz" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.782625 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.783490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-config-data\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.786940 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vlg4\" (UniqueName: \"kubernetes.io/projected/596ec651-2088-4459-b0dd-224c67150b5d-kube-api-access-2vlg4\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.788163 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-scripts\") pod \"ceilometer-0\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " pod="openstack/ceilometer-0" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.808328 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.820674 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:46 crc kubenswrapper[4809]: I1127 17:38:46.830605 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.086142 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.090068 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.100363 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.100823 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.101034 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bh9xs" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.113104 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.184818 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.186823 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.196641 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.198915 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.288235 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lv8h9"] Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291092 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291148 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291181 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291209 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-config-data\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291250 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291271 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291340 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-scripts\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-logs\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291459 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291491 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291558 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlcng\" (UniqueName: \"kubernetes.io/projected/3acb572b-b007-4b06-9efb-3334da96a7dc-kube-api-access-vlcng\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291596 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291642 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-schwb\" (UniqueName: \"kubernetes.io/projected/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-kube-api-access-schwb\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.291725 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.368374 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9"] Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.395915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.395967 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.395983 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-scripts\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.395999 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-logs\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396081 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlcng\" (UniqueName: \"kubernetes.io/projected/3acb572b-b007-4b06-9efb-3334da96a7dc-kube-api-access-vlcng\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396103 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396128 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-schwb\" (UniqueName: \"kubernetes.io/projected/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-kube-api-access-schwb\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396161 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396245 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396266 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396282 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.396298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-config-data\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.398218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.398559 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-logs\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.398601 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.398901 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.399157 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.402869 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.413492 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.413669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-scripts\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.415611 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-config-data\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.418086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.419862 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlcng\" (UniqueName: \"kubernetes.io/projected/3acb572b-b007-4b06-9efb-3334da96a7dc-kube-api-access-vlcng\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.420103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.458596 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-schwb\" (UniqueName: \"kubernetes.io/projected/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-kube-api-access-schwb\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.468345 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.468510 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.487298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.538536 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.546411 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.606474 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-sb\") pod \"ab62f250-6ba2-4c46-bda1-061a52f9928a\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.606558 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-config\") pod \"ab62f250-6ba2-4c46-bda1-061a52f9928a\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.606626 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-svc\") pod \"ab62f250-6ba2-4c46-bda1-061a52f9928a\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.606766 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j24gm\" (UniqueName: \"kubernetes.io/projected/ab62f250-6ba2-4c46-bda1-061a52f9928a-kube-api-access-j24gm\") pod \"ab62f250-6ba2-4c46-bda1-061a52f9928a\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.606782 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-swift-storage-0\") pod \"ab62f250-6ba2-4c46-bda1-061a52f9928a\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.606853 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-nb\") pod \"ab62f250-6ba2-4c46-bda1-061a52f9928a\" (UID: \"ab62f250-6ba2-4c46-bda1-061a52f9928a\") " Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.670009 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab62f250-6ba2-4c46-bda1-061a52f9928a-kube-api-access-j24gm" (OuterVolumeSpecName: "kube-api-access-j24gm") pod "ab62f250-6ba2-4c46-bda1-061a52f9928a" (UID: "ab62f250-6ba2-4c46-bda1-061a52f9928a"). InnerVolumeSpecName "kube-api-access-j24gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.670435 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-t2gnx"] Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.682565 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-97thg"] Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.708753 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j24gm\" (UniqueName: \"kubernetes.io/projected/ab62f250-6ba2-4c46-bda1-061a52f9928a-kube-api-access-j24gm\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.746815 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.792797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" event={"ID":"b91aa6be-d5c9-4342-80b5-69011df32c42","Type":"ContainerStarted","Data":"00b610f8f18c3554d127a02000be605cc944ab58abeb204d49819d6cfe37b8c3"} Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.804785 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lv8h9" event={"ID":"ae4b6101-b917-4224-bc83-0b8e9327013c","Type":"ContainerStarted","Data":"3749c74e9a5ac43b5b436c933999e401cf29ff7f12f499db3d55ed20264316a1"} Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.814123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-t2gnx" event={"ID":"9a3b3964-4402-49a4-86ba-e6e7909082ec","Type":"ContainerStarted","Data":"e76ed9602e46c1d8479bb657c98a72e04515d23f7b25b2ef5ec8292ec733365d"} Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.830499 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-97thg" event={"ID":"c184a0c3-5bb0-49f0-966f-fb565fd13202","Type":"ContainerStarted","Data":"d074b23df3560cd4234217ac9340a476f2216b05af00993aa1a9311608f6d591"} Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.846833 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" event={"ID":"ab62f250-6ba2-4c46-bda1-061a52f9928a","Type":"ContainerDied","Data":"6c54205c21810dfdcf8581df46a6d8215bfa7ecbe2bfc4d242f751abe958d54a"} Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.846887 4809 scope.go:117] "RemoveContainer" containerID="14d3407f4fca243a7afec7ae5ed60078840ad3ffbecfed5804c64ed4362eee63" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.847018 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-jhdt8" Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.868102 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-n7gbr"] Nov 27 17:38:47 crc kubenswrapper[4809]: I1127 17:38:47.877188 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:38:47 crc kubenswrapper[4809]: W1127 17:38:47.923535 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod596ec651_2088_4459_b0dd_224c67150b5d.slice/crio-587b6cf5aaed97ed37e874602721ffa02025c9d285cf1a12d87b4b1af6fd1ac1 WatchSource:0}: Error finding container 587b6cf5aaed97ed37e874602721ffa02025c9d285cf1a12d87b4b1af6fd1ac1: Status 404 returned error can't find the container with id 587b6cf5aaed97ed37e874602721ffa02025c9d285cf1a12d87b4b1af6fd1ac1 Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.058158 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ab62f250-6ba2-4c46-bda1-061a52f9928a" (UID: "ab62f250-6ba2-4c46-bda1-061a52f9928a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.061163 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-rhmfz"] Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.062411 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ab62f250-6ba2-4c46-bda1-061a52f9928a" (UID: "ab62f250-6ba2-4c46-bda1-061a52f9928a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.079593 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ktqz6"] Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.114006 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ab62f250-6ba2-4c46-bda1-061a52f9928a" (UID: "ab62f250-6ba2-4c46-bda1-061a52f9928a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.121118 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.121140 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.121151 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.143039 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ab62f250-6ba2-4c46-bda1-061a52f9928a" (UID: "ab62f250-6ba2-4c46-bda1-061a52f9928a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.196977 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-config" (OuterVolumeSpecName: "config") pod "ab62f250-6ba2-4c46-bda1-061a52f9928a" (UID: "ab62f250-6ba2-4c46-bda1-061a52f9928a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.223138 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.223188 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab62f250-6ba2-4c46-bda1-061a52f9928a-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.294550 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:38:48 crc kubenswrapper[4809]: W1127 17:38:48.312689 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5438552_cd8f_43f1_b478_fb63f2f1e6c5.slice/crio-aa3d88b5a767042bdc07171af2e7b6a53a64b66462541361e9b12497f25878b7 WatchSource:0}: Error finding container aa3d88b5a767042bdc07171af2e7b6a53a64b66462541361e9b12497f25878b7: Status 404 returned error can't find the container with id aa3d88b5a767042bdc07171af2e7b6a53a64b66462541361e9b12497f25878b7 Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.511882 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:38:48 crc kubenswrapper[4809]: W1127 17:38:48.558452 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3acb572b_b007_4b06_9efb_3334da96a7dc.slice/crio-933edae32b381576f6ef0429cdf86acb602f26f8bff4ed38ef02369ef2091bbd WatchSource:0}: Error finding container 933edae32b381576f6ef0429cdf86acb602f26f8bff4ed38ef02369ef2091bbd: Status 404 returned error can't find the container with id 933edae32b381576f6ef0429cdf86acb602f26f8bff4ed38ef02369ef2091bbd Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.584947 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jhdt8"] Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.592796 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jhdt8"] Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.862123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5438552-cd8f-43f1-b478-fb63f2f1e6c5","Type":"ContainerStarted","Data":"aa3d88b5a767042bdc07171af2e7b6a53a64b66462541361e9b12497f25878b7"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.864685 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-97thg" event={"ID":"c184a0c3-5bb0-49f0-966f-fb565fd13202","Type":"ContainerStarted","Data":"6de46e7d9a5928fd2d465277012c3423b9e297c739d698796bfebc205c6b9091"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.880104 4809 generic.go:334] "Generic (PLEG): container finished" podID="b91aa6be-d5c9-4342-80b5-69011df32c42" containerID="67c07faaf494277dee2b6b5406435530f0f77824ba8e6d664df46da0e5dca48d" exitCode=0 Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.880190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" event={"ID":"b91aa6be-d5c9-4342-80b5-69011df32c42","Type":"ContainerDied","Data":"67c07faaf494277dee2b6b5406435530f0f77824ba8e6d664df46da0e5dca48d"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.887520 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-97thg" podStartSLOduration=2.887501274 podStartE2EDuration="2.887501274s" podCreationTimestamp="2025-11-27 17:38:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:48.881284065 +0000 UTC m=+1764.153741417" watchObservedRunningTime="2025-11-27 17:38:48.887501274 +0000 UTC m=+1764.159958626" Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.906261 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lv8h9" event={"ID":"ae4b6101-b917-4224-bc83-0b8e9327013c","Type":"ContainerStarted","Data":"c8fb2111f80e8c44f310c5346c2a3ee96da1beaf0221cf3591832a76006b7ad0"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.915311 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3acb572b-b007-4b06-9efb-3334da96a7dc","Type":"ContainerStarted","Data":"933edae32b381576f6ef0429cdf86acb602f26f8bff4ed38ef02369ef2091bbd"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.925035 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ktqz6" event={"ID":"57b62f0e-7576-4e57-9993-30000f0eee5a","Type":"ContainerStarted","Data":"aa8c7ebb5170f3af85b5c2194ff1184ca2726e75149b6dd707629693f84d927d"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.927282 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596ec651-2088-4459-b0dd-224c67150b5d","Type":"ContainerStarted","Data":"587b6cf5aaed97ed37e874602721ffa02025c9d285cf1a12d87b4b1af6fd1ac1"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.933335 4809 generic.go:334] "Generic (PLEG): container finished" podID="51a633d0-f7b6-4b1c-b609-7f9684069609" containerID="88a26ef4535bdf518029354840c0f3655cdf3e182a723f2311029144ba64dfba" exitCode=0 Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.933396 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" event={"ID":"51a633d0-f7b6-4b1c-b609-7f9684069609","Type":"ContainerDied","Data":"88a26ef4535bdf518029354840c0f3655cdf3e182a723f2311029144ba64dfba"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.933420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" event={"ID":"51a633d0-f7b6-4b1c-b609-7f9684069609","Type":"ContainerStarted","Data":"f0e4f95c0d20993839a5eb37c88908770ed6c65a5fe317d199d4b590a345d907"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.937025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-rhmfz" event={"ID":"4e177c61-cc3e-444a-b6ed-aecd6d262b7c","Type":"ContainerStarted","Data":"cf04a7d3faa3280057a00eab61b82df001134464cb7cf98bec14a4048794a7bc"} Nov 27 17:38:48 crc kubenswrapper[4809]: I1127 17:38:48.942768 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-lv8h9" podStartSLOduration=3.9427549280000003 podStartE2EDuration="3.942754928s" podCreationTimestamp="2025-11-27 17:38:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:48.941627148 +0000 UTC m=+1764.214084500" watchObservedRunningTime="2025-11-27 17:38:48.942754928 +0000 UTC m=+1764.215212280" Nov 27 17:38:49 crc kubenswrapper[4809]: I1127 17:38:49.477984 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab62f250-6ba2-4c46-bda1-061a52f9928a" path="/var/lib/kubelet/pods/ab62f250-6ba2-4c46-bda1-061a52f9928a/volumes" Nov 27 17:38:49 crc kubenswrapper[4809]: I1127 17:38:49.779091 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:38:49 crc kubenswrapper[4809]: I1127 17:38:49.880795 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:38:49 crc kubenswrapper[4809]: I1127 17:38:49.896284 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:38:49 crc kubenswrapper[4809]: I1127 17:38:49.968107 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" event={"ID":"51a633d0-f7b6-4b1c-b609-7f9684069609","Type":"ContainerStarted","Data":"aec02d90c49dd7261e47b119faa8e613081399e8e11664be13b98bf493f6f61b"} Nov 27 17:38:49 crc kubenswrapper[4809]: I1127 17:38:49.969234 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:49 crc kubenswrapper[4809]: I1127 17:38:49.981377 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5438552-cd8f-43f1-b478-fb63f2f1e6c5","Type":"ContainerStarted","Data":"6ba7ea18d3173de854a5a126b640ee54a4b0a6069ebeda3fc63d2211da46d004"} Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.002712 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" podStartSLOduration=4.002695247 podStartE2EDuration="4.002695247s" podCreationTimestamp="2025-11-27 17:38:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:50.001151616 +0000 UTC m=+1765.273608988" watchObservedRunningTime="2025-11-27 17:38:50.002695247 +0000 UTC m=+1765.275152599" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.009559 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3acb572b-b007-4b06-9efb-3334da96a7dc","Type":"ContainerStarted","Data":"dcc22cee3660cf4dcf003c5732d955c2ffc643b21f790e70ed0c67f8af01bea0"} Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.458219 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:38:50 crc kubenswrapper[4809]: E1127 17:38:50.460062 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.482614 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.605808 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-swift-storage-0\") pod \"b91aa6be-d5c9-4342-80b5-69011df32c42\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.605871 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-sb\") pod \"b91aa6be-d5c9-4342-80b5-69011df32c42\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.605974 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-svc\") pod \"b91aa6be-d5c9-4342-80b5-69011df32c42\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.606025 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4746t\" (UniqueName: \"kubernetes.io/projected/b91aa6be-d5c9-4342-80b5-69011df32c42-kube-api-access-4746t\") pod \"b91aa6be-d5c9-4342-80b5-69011df32c42\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.606139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-nb\") pod \"b91aa6be-d5c9-4342-80b5-69011df32c42\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.606226 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-config\") pod \"b91aa6be-d5c9-4342-80b5-69011df32c42\" (UID: \"b91aa6be-d5c9-4342-80b5-69011df32c42\") " Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.631696 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b91aa6be-d5c9-4342-80b5-69011df32c42-kube-api-access-4746t" (OuterVolumeSpecName: "kube-api-access-4746t") pod "b91aa6be-d5c9-4342-80b5-69011df32c42" (UID: "b91aa6be-d5c9-4342-80b5-69011df32c42"). InnerVolumeSpecName "kube-api-access-4746t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.634404 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b91aa6be-d5c9-4342-80b5-69011df32c42" (UID: "b91aa6be-d5c9-4342-80b5-69011df32c42"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.636945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b91aa6be-d5c9-4342-80b5-69011df32c42" (UID: "b91aa6be-d5c9-4342-80b5-69011df32c42"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.638797 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b91aa6be-d5c9-4342-80b5-69011df32c42" (UID: "b91aa6be-d5c9-4342-80b5-69011df32c42"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.655683 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b91aa6be-d5c9-4342-80b5-69011df32c42" (UID: "b91aa6be-d5c9-4342-80b5-69011df32c42"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.655832 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-config" (OuterVolumeSpecName: "config") pod "b91aa6be-d5c9-4342-80b5-69011df32c42" (UID: "b91aa6be-d5c9-4342-80b5-69011df32c42"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.709220 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.709273 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.709285 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.709293 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4746t\" (UniqueName: \"kubernetes.io/projected/b91aa6be-d5c9-4342-80b5-69011df32c42-kube-api-access-4746t\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.709368 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:50 crc kubenswrapper[4809]: I1127 17:38:50.709383 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91aa6be-d5c9-4342-80b5-69011df32c42-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.022726 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" event={"ID":"b91aa6be-d5c9-4342-80b5-69011df32c42","Type":"ContainerDied","Data":"00b610f8f18c3554d127a02000be605cc944ab58abeb204d49819d6cfe37b8c3"} Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.022878 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9" Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.023126 4809 scope.go:117] "RemoveContainer" containerID="67c07faaf494277dee2b6b5406435530f0f77824ba8e6d664df46da0e5dca48d" Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.035154 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5438552-cd8f-43f1-b478-fb63f2f1e6c5","Type":"ContainerStarted","Data":"9c6672bdbbe432d81c4b9870281a616352852256521ac3b6724e7b33e4d666d6"} Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.035246 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerName="glance-log" containerID="cri-o://6ba7ea18d3173de854a5a126b640ee54a4b0a6069ebeda3fc63d2211da46d004" gracePeriod=30 Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.035319 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerName="glance-httpd" containerID="cri-o://9c6672bdbbe432d81c4b9870281a616352852256521ac3b6724e7b33e4d666d6" gracePeriod=30 Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.037695 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3acb572b-b007-4b06-9efb-3334da96a7dc","Type":"ContainerStarted","Data":"c00e7ba540e0c237ec2c605e13b3126c9a1a709b0dd06afe3f123acb9c2c7068"} Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.037883 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerName="glance-log" containerID="cri-o://dcc22cee3660cf4dcf003c5732d955c2ffc643b21f790e70ed0c67f8af01bea0" gracePeriod=30 Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.037904 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerName="glance-httpd" containerID="cri-o://c00e7ba540e0c237ec2c605e13b3126c9a1a709b0dd06afe3f123acb9c2c7068" gracePeriod=30 Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.063655 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.063616634 podStartE2EDuration="5.063616634s" podCreationTimestamp="2025-11-27 17:38:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:51.056087828 +0000 UTC m=+1766.328545190" watchObservedRunningTime="2025-11-27 17:38:51.063616634 +0000 UTC m=+1766.336073996" Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.099827 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.099804809 podStartE2EDuration="5.099804809s" podCreationTimestamp="2025-11-27 17:38:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:38:51.093992771 +0000 UTC m=+1766.366450133" watchObservedRunningTime="2025-11-27 17:38:51.099804809 +0000 UTC m=+1766.372262161" Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.172203 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9"] Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.189750 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-mwtf9"] Nov 27 17:38:51 crc kubenswrapper[4809]: I1127 17:38:51.477040 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b91aa6be-d5c9-4342-80b5-69011df32c42" path="/var/lib/kubelet/pods/b91aa6be-d5c9-4342-80b5-69011df32c42/volumes" Nov 27 17:38:52 crc kubenswrapper[4809]: I1127 17:38:52.050549 4809 generic.go:334] "Generic (PLEG): container finished" podID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerID="9c6672bdbbe432d81c4b9870281a616352852256521ac3b6724e7b33e4d666d6" exitCode=0 Nov 27 17:38:52 crc kubenswrapper[4809]: I1127 17:38:52.050602 4809 generic.go:334] "Generic (PLEG): container finished" podID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerID="6ba7ea18d3173de854a5a126b640ee54a4b0a6069ebeda3fc63d2211da46d004" exitCode=143 Nov 27 17:38:52 crc kubenswrapper[4809]: I1127 17:38:52.050604 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5438552-cd8f-43f1-b478-fb63f2f1e6c5","Type":"ContainerDied","Data":"9c6672bdbbe432d81c4b9870281a616352852256521ac3b6724e7b33e4d666d6"} Nov 27 17:38:52 crc kubenswrapper[4809]: I1127 17:38:52.050642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5438552-cd8f-43f1-b478-fb63f2f1e6c5","Type":"ContainerDied","Data":"6ba7ea18d3173de854a5a126b640ee54a4b0a6069ebeda3fc63d2211da46d004"} Nov 27 17:38:52 crc kubenswrapper[4809]: I1127 17:38:52.054716 4809 generic.go:334] "Generic (PLEG): container finished" podID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerID="c00e7ba540e0c237ec2c605e13b3126c9a1a709b0dd06afe3f123acb9c2c7068" exitCode=0 Nov 27 17:38:52 crc kubenswrapper[4809]: I1127 17:38:52.054757 4809 generic.go:334] "Generic (PLEG): container finished" podID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerID="dcc22cee3660cf4dcf003c5732d955c2ffc643b21f790e70ed0c67f8af01bea0" exitCode=143 Nov 27 17:38:52 crc kubenswrapper[4809]: I1127 17:38:52.054777 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3acb572b-b007-4b06-9efb-3334da96a7dc","Type":"ContainerDied","Data":"c00e7ba540e0c237ec2c605e13b3126c9a1a709b0dd06afe3f123acb9c2c7068"} Nov 27 17:38:52 crc kubenswrapper[4809]: I1127 17:38:52.054826 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3acb572b-b007-4b06-9efb-3334da96a7dc","Type":"ContainerDied","Data":"dcc22cee3660cf4dcf003c5732d955c2ffc643b21f790e70ed0c67f8af01bea0"} Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.067981 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae4b6101-b917-4224-bc83-0b8e9327013c" containerID="c8fb2111f80e8c44f310c5346c2a3ee96da1beaf0221cf3591832a76006b7ad0" exitCode=0 Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.068326 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lv8h9" event={"ID":"ae4b6101-b917-4224-bc83-0b8e9327013c","Type":"ContainerDied","Data":"c8fb2111f80e8c44f310c5346c2a3ee96da1beaf0221cf3591832a76006b7ad0"} Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.637323 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.645680 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771368 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-config-data\") pod \"3acb572b-b007-4b06-9efb-3334da96a7dc\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771499 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-combined-ca-bundle\") pod \"3acb572b-b007-4b06-9efb-3334da96a7dc\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771534 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771581 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-combined-ca-bundle\") pod \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771612 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-scripts\") pod \"3acb572b-b007-4b06-9efb-3334da96a7dc\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771690 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-httpd-run\") pod \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771786 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"3acb572b-b007-4b06-9efb-3334da96a7dc\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771812 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-httpd-run\") pod \"3acb572b-b007-4b06-9efb-3334da96a7dc\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771838 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-scripts\") pod \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771865 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlcng\" (UniqueName: \"kubernetes.io/projected/3acb572b-b007-4b06-9efb-3334da96a7dc-kube-api-access-vlcng\") pod \"3acb572b-b007-4b06-9efb-3334da96a7dc\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771887 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-logs\") pod \"3acb572b-b007-4b06-9efb-3334da96a7dc\" (UID: \"3acb572b-b007-4b06-9efb-3334da96a7dc\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771919 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-schwb\" (UniqueName: \"kubernetes.io/projected/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-kube-api-access-schwb\") pod \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771946 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-config-data\") pod \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.771970 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-logs\") pod \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\" (UID: \"f5438552-cd8f-43f1-b478-fb63f2f1e6c5\") " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.772799 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-logs" (OuterVolumeSpecName: "logs") pod "3acb572b-b007-4b06-9efb-3334da96a7dc" (UID: "3acb572b-b007-4b06-9efb-3334da96a7dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.772964 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-logs" (OuterVolumeSpecName: "logs") pod "f5438552-cd8f-43f1-b478-fb63f2f1e6c5" (UID: "f5438552-cd8f-43f1-b478-fb63f2f1e6c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.773612 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3acb572b-b007-4b06-9efb-3334da96a7dc" (UID: "3acb572b-b007-4b06-9efb-3334da96a7dc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.773685 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f5438552-cd8f-43f1-b478-fb63f2f1e6c5" (UID: "f5438552-cd8f-43f1-b478-fb63f2f1e6c5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.780340 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3acb572b-b007-4b06-9efb-3334da96a7dc-kube-api-access-vlcng" (OuterVolumeSpecName: "kube-api-access-vlcng") pod "3acb572b-b007-4b06-9efb-3334da96a7dc" (UID: "3acb572b-b007-4b06-9efb-3334da96a7dc"). InnerVolumeSpecName "kube-api-access-vlcng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.781540 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "3acb572b-b007-4b06-9efb-3334da96a7dc" (UID: "3acb572b-b007-4b06-9efb-3334da96a7dc"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.782107 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-scripts" (OuterVolumeSpecName: "scripts") pod "f5438552-cd8f-43f1-b478-fb63f2f1e6c5" (UID: "f5438552-cd8f-43f1-b478-fb63f2f1e6c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.782207 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "f5438552-cd8f-43f1-b478-fb63f2f1e6c5" (UID: "f5438552-cd8f-43f1-b478-fb63f2f1e6c5"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.782250 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-kube-api-access-schwb" (OuterVolumeSpecName: "kube-api-access-schwb") pod "f5438552-cd8f-43f1-b478-fb63f2f1e6c5" (UID: "f5438552-cd8f-43f1-b478-fb63f2f1e6c5"). InnerVolumeSpecName "kube-api-access-schwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.782299 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-scripts" (OuterVolumeSpecName: "scripts") pod "3acb572b-b007-4b06-9efb-3334da96a7dc" (UID: "3acb572b-b007-4b06-9efb-3334da96a7dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.826122 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5438552-cd8f-43f1-b478-fb63f2f1e6c5" (UID: "f5438552-cd8f-43f1-b478-fb63f2f1e6c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.827203 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3acb572b-b007-4b06-9efb-3334da96a7dc" (UID: "3acb572b-b007-4b06-9efb-3334da96a7dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.833801 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-config-data" (OuterVolumeSpecName: "config-data") pod "f5438552-cd8f-43f1-b478-fb63f2f1e6c5" (UID: "f5438552-cd8f-43f1-b478-fb63f2f1e6c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.840546 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-config-data" (OuterVolumeSpecName: "config-data") pod "3acb572b-b007-4b06-9efb-3334da96a7dc" (UID: "3acb572b-b007-4b06-9efb-3334da96a7dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876080 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876206 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876268 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876287 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876302 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3acb572b-b007-4b06-9efb-3334da96a7dc-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876316 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876349 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876362 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876374 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876386 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlcng\" (UniqueName: \"kubernetes.io/projected/3acb572b-b007-4b06-9efb-3334da96a7dc-kube-api-access-vlcng\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876401 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3acb572b-b007-4b06-9efb-3334da96a7dc-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876413 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-schwb\" (UniqueName: \"kubernetes.io/projected/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-kube-api-access-schwb\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876425 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.876436 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5438552-cd8f-43f1-b478-fb63f2f1e6c5-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.895578 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.904244 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.979592 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:53 crc kubenswrapper[4809]: I1127 17:38:53.979626 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.081413 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3acb572b-b007-4b06-9efb-3334da96a7dc","Type":"ContainerDied","Data":"933edae32b381576f6ef0429cdf86acb602f26f8bff4ed38ef02369ef2091bbd"} Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.081461 4809 scope.go:117] "RemoveContainer" containerID="c00e7ba540e0c237ec2c605e13b3126c9a1a709b0dd06afe3f123acb9c2c7068" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.081426 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.084188 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.084316 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5438552-cd8f-43f1-b478-fb63f2f1e6c5","Type":"ContainerDied","Data":"aa3d88b5a767042bdc07171af2e7b6a53a64b66462541361e9b12497f25878b7"} Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.156037 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.163049 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.181938 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:38:54 crc kubenswrapper[4809]: E1127 17:38:54.182935 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerName="glance-httpd" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.182957 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerName="glance-httpd" Nov 27 17:38:54 crc kubenswrapper[4809]: E1127 17:38:54.182988 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerName="glance-log" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.182995 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerName="glance-log" Nov 27 17:38:54 crc kubenswrapper[4809]: E1127 17:38:54.183014 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab62f250-6ba2-4c46-bda1-061a52f9928a" containerName="init" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183022 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab62f250-6ba2-4c46-bda1-061a52f9928a" containerName="init" Nov 27 17:38:54 crc kubenswrapper[4809]: E1127 17:38:54.183047 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerName="glance-log" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183057 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerName="glance-log" Nov 27 17:38:54 crc kubenswrapper[4809]: E1127 17:38:54.183071 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerName="glance-httpd" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183078 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerName="glance-httpd" Nov 27 17:38:54 crc kubenswrapper[4809]: E1127 17:38:54.183098 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b91aa6be-d5c9-4342-80b5-69011df32c42" containerName="init" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183105 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b91aa6be-d5c9-4342-80b5-69011df32c42" containerName="init" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183490 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerName="glance-httpd" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183515 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b91aa6be-d5c9-4342-80b5-69011df32c42" containerName="init" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183542 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" containerName="glance-log" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183566 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab62f250-6ba2-4c46-bda1-061a52f9928a" containerName="init" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183583 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerName="glance-log" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.183592 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3acb572b-b007-4b06-9efb-3334da96a7dc" containerName="glance-httpd" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.185875 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.198308 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.199288 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.199577 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bh9xs" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.236663 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.256433 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.267190 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.279671 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.281702 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.284355 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.287670 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.287724 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.287773 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.287830 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.287959 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgzvx\" (UniqueName: \"kubernetes.io/projected/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-kube-api-access-vgzvx\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.288112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.288159 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.310449 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389279 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389613 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389644 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389694 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-scripts\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389722 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389811 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389839 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-logs\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389870 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-config-data\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389912 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqjg8\" (UniqueName: \"kubernetes.io/projected/16ea53cb-7781-4900-adaa-29ac80ee734d-kube-api-access-wqjg8\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389950 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgzvx\" (UniqueName: \"kubernetes.io/projected/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-kube-api-access-vgzvx\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.389968 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.390347 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.391242 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.392488 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.395003 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.396118 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.412402 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.430513 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgzvx\" (UniqueName: \"kubernetes.io/projected/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-kube-api-access-vgzvx\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.451850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.492856 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.492922 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.492950 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-logs\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.492993 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.493028 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-config-data\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.493047 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqjg8\" (UniqueName: \"kubernetes.io/projected/16ea53cb-7781-4900-adaa-29ac80ee734d-kube-api-access-wqjg8\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.493145 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-scripts\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.493911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-logs\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.494949 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.496972 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.504198 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-scripts\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.504830 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.530778 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.531387 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-config-data\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.535410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqjg8\" (UniqueName: \"kubernetes.io/projected/16ea53cb-7781-4900-adaa-29ac80ee734d-kube-api-access-wqjg8\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.569807 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " pod="openstack/glance-default-external-api-0" Nov 27 17:38:54 crc kubenswrapper[4809]: I1127 17:38:54.602772 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:38:55 crc kubenswrapper[4809]: I1127 17:38:55.170523 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:38:55 crc kubenswrapper[4809]: I1127 17:38:55.270938 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:38:55 crc kubenswrapper[4809]: I1127 17:38:55.470326 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3acb572b-b007-4b06-9efb-3334da96a7dc" path="/var/lib/kubelet/pods/3acb572b-b007-4b06-9efb-3334da96a7dc/volumes" Nov 27 17:38:55 crc kubenswrapper[4809]: I1127 17:38:55.471173 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5438552-cd8f-43f1-b478-fb63f2f1e6c5" path="/var/lib/kubelet/pods/f5438552-cd8f-43f1-b478-fb63f2f1e6c5/volumes" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.529562 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.634565 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89bz8\" (UniqueName: \"kubernetes.io/projected/ae4b6101-b917-4224-bc83-0b8e9327013c-kube-api-access-89bz8\") pod \"ae4b6101-b917-4224-bc83-0b8e9327013c\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.634686 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-combined-ca-bundle\") pod \"ae4b6101-b917-4224-bc83-0b8e9327013c\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.634723 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-fernet-keys\") pod \"ae4b6101-b917-4224-bc83-0b8e9327013c\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.634793 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-credential-keys\") pod \"ae4b6101-b917-4224-bc83-0b8e9327013c\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.634955 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-scripts\") pod \"ae4b6101-b917-4224-bc83-0b8e9327013c\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.634986 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-config-data\") pod \"ae4b6101-b917-4224-bc83-0b8e9327013c\" (UID: \"ae4b6101-b917-4224-bc83-0b8e9327013c\") " Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.643660 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae4b6101-b917-4224-bc83-0b8e9327013c-kube-api-access-89bz8" (OuterVolumeSpecName: "kube-api-access-89bz8") pod "ae4b6101-b917-4224-bc83-0b8e9327013c" (UID: "ae4b6101-b917-4224-bc83-0b8e9327013c"). InnerVolumeSpecName "kube-api-access-89bz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.643923 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ae4b6101-b917-4224-bc83-0b8e9327013c" (UID: "ae4b6101-b917-4224-bc83-0b8e9327013c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.650818 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-scripts" (OuterVolumeSpecName: "scripts") pod "ae4b6101-b917-4224-bc83-0b8e9327013c" (UID: "ae4b6101-b917-4224-bc83-0b8e9327013c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.652615 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ae4b6101-b917-4224-bc83-0b8e9327013c" (UID: "ae4b6101-b917-4224-bc83-0b8e9327013c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.671688 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae4b6101-b917-4224-bc83-0b8e9327013c" (UID: "ae4b6101-b917-4224-bc83-0b8e9327013c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.671905 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-config-data" (OuterVolumeSpecName: "config-data") pod "ae4b6101-b917-4224-bc83-0b8e9327013c" (UID: "ae4b6101-b917-4224-bc83-0b8e9327013c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.737428 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.737466 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.737481 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89bz8\" (UniqueName: \"kubernetes.io/projected/ae4b6101-b917-4224-bc83-0b8e9327013c-kube-api-access-89bz8\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.737498 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.737509 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.737519 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4b6101-b917-4224-bc83-0b8e9327013c-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.823327 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.890444 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-w5bp6"] Nov 27 17:38:56 crc kubenswrapper[4809]: I1127 17:38:56.890708 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" podUID="d542ca5d-c018-4f64-96f0-e515d148c51c" containerName="dnsmasq-dns" containerID="cri-o://76cdd523a6e9bc63ffd51c4fb46910fc4a38ea7c8f838aca7f6f6ccf855f3452" gracePeriod=10 Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.173447 4809 generic.go:334] "Generic (PLEG): container finished" podID="d542ca5d-c018-4f64-96f0-e515d148c51c" containerID="76cdd523a6e9bc63ffd51c4fb46910fc4a38ea7c8f838aca7f6f6ccf855f3452" exitCode=0 Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.173856 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" event={"ID":"d542ca5d-c018-4f64-96f0-e515d148c51c","Type":"ContainerDied","Data":"76cdd523a6e9bc63ffd51c4fb46910fc4a38ea7c8f838aca7f6f6ccf855f3452"} Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.195445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lv8h9" event={"ID":"ae4b6101-b917-4224-bc83-0b8e9327013c","Type":"ContainerDied","Data":"3749c74e9a5ac43b5b436c933999e401cf29ff7f12f499db3d55ed20264316a1"} Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.195488 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3749c74e9a5ac43b5b436c933999e401cf29ff7f12f499db3d55ed20264316a1" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.195560 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lv8h9" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.616178 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-lv8h9"] Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.623924 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-lv8h9"] Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.713032 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-k747n"] Nov 27 17:38:57 crc kubenswrapper[4809]: E1127 17:38:57.713455 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae4b6101-b917-4224-bc83-0b8e9327013c" containerName="keystone-bootstrap" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.713480 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae4b6101-b917-4224-bc83-0b8e9327013c" containerName="keystone-bootstrap" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.713759 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae4b6101-b917-4224-bc83-0b8e9327013c" containerName="keystone-bootstrap" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.714540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.718230 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-l4mwz" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.718422 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.718630 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.719213 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.723608 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.729938 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-k747n"] Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.870651 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-fernet-keys\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.870774 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-credential-keys\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.870800 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2fbt\" (UniqueName: \"kubernetes.io/projected/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-kube-api-access-l2fbt\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.871712 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-config-data\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.871833 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-combined-ca-bundle\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.871900 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-scripts\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.974025 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-fernet-keys\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.974148 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-credential-keys\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.974174 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2fbt\" (UniqueName: \"kubernetes.io/projected/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-kube-api-access-l2fbt\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.974211 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-config-data\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.974257 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-combined-ca-bundle\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.974289 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-scripts\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.980342 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-credential-keys\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.980601 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-fernet-keys\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.980798 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-combined-ca-bundle\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.982003 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-scripts\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.989638 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-config-data\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:57 crc kubenswrapper[4809]: I1127 17:38:57.993187 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2fbt\" (UniqueName: \"kubernetes.io/projected/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-kube-api-access-l2fbt\") pod \"keystone-bootstrap-k747n\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:58 crc kubenswrapper[4809]: I1127 17:38:58.045061 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k747n" Nov 27 17:38:59 crc kubenswrapper[4809]: I1127 17:38:59.471341 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae4b6101-b917-4224-bc83-0b8e9327013c" path="/var/lib/kubelet/pods/ae4b6101-b917-4224-bc83-0b8e9327013c/volumes" Nov 27 17:38:59 crc kubenswrapper[4809]: I1127 17:38:59.970355 4809 scope.go:117] "RemoveContainer" containerID="dcc22cee3660cf4dcf003c5732d955c2ffc643b21f790e70ed0c67f8af01bea0" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.080041 4809 scope.go:117] "RemoveContainer" containerID="9c6672bdbbe432d81c4b9870281a616352852256521ac3b6724e7b33e4d666d6" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.341807 4809 scope.go:117] "RemoveContainer" containerID="6ba7ea18d3173de854a5a126b640ee54a4b0a6069ebeda3fc63d2211da46d004" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.480844 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:39:00 crc kubenswrapper[4809]: E1127 17:39:00.481116 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.558632 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.636838 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-config\") pod \"d542ca5d-c018-4f64-96f0-e515d148c51c\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.637002 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-sb\") pod \"d542ca5d-c018-4f64-96f0-e515d148c51c\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.637043 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qkmx\" (UniqueName: \"kubernetes.io/projected/d542ca5d-c018-4f64-96f0-e515d148c51c-kube-api-access-5qkmx\") pod \"d542ca5d-c018-4f64-96f0-e515d148c51c\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.637105 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-nb\") pod \"d542ca5d-c018-4f64-96f0-e515d148c51c\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.637137 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-swift-storage-0\") pod \"d542ca5d-c018-4f64-96f0-e515d148c51c\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.637231 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-svc\") pod \"d542ca5d-c018-4f64-96f0-e515d148c51c\" (UID: \"d542ca5d-c018-4f64-96f0-e515d148c51c\") " Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.648716 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d542ca5d-c018-4f64-96f0-e515d148c51c-kube-api-access-5qkmx" (OuterVolumeSpecName: "kube-api-access-5qkmx") pod "d542ca5d-c018-4f64-96f0-e515d148c51c" (UID: "d542ca5d-c018-4f64-96f0-e515d148c51c"). InnerVolumeSpecName "kube-api-access-5qkmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.702484 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d542ca5d-c018-4f64-96f0-e515d148c51c" (UID: "d542ca5d-c018-4f64-96f0-e515d148c51c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.707658 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-config" (OuterVolumeSpecName: "config") pod "d542ca5d-c018-4f64-96f0-e515d148c51c" (UID: "d542ca5d-c018-4f64-96f0-e515d148c51c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.721176 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d542ca5d-c018-4f64-96f0-e515d148c51c" (UID: "d542ca5d-c018-4f64-96f0-e515d148c51c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.732633 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d542ca5d-c018-4f64-96f0-e515d148c51c" (UID: "d542ca5d-c018-4f64-96f0-e515d148c51c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.739013 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.739043 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qkmx\" (UniqueName: \"kubernetes.io/projected/d542ca5d-c018-4f64-96f0-e515d148c51c-kube-api-access-5qkmx\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.739058 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.739067 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.739076 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.756625 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d542ca5d-c018-4f64-96f0-e515d148c51c" (UID: "d542ca5d-c018-4f64-96f0-e515d148c51c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.840708 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d542ca5d-c018-4f64-96f0-e515d148c51c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.866085 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.985007 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:39:00 crc kubenswrapper[4809]: I1127 17:39:00.995117 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-k747n"] Nov 27 17:39:01 crc kubenswrapper[4809]: W1127 17:39:01.003308 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bbb5fa1_67a7_4d6e_9354_36598e97d9d0.slice/crio-1cf72c60523e5210fd753ed23b3a004850d4960854894d77561575cc839a34d8 WatchSource:0}: Error finding container 1cf72c60523e5210fd753ed23b3a004850d4960854894d77561575cc839a34d8: Status 404 returned error can't find the container with id 1cf72c60523e5210fd753ed23b3a004850d4960854894d77561575cc839a34d8 Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.337377 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-rhmfz" event={"ID":"4e177c61-cc3e-444a-b6ed-aecd6d262b7c","Type":"ContainerStarted","Data":"add867ab2e890587644a3fa7893d6444487f2b07f654196452bfdcc21c669ea8"} Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.341030 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16ea53cb-7781-4900-adaa-29ac80ee734d","Type":"ContainerStarted","Data":"4a7efe5ca33543e6109f6bab0bbd220d4db651bdfcac38f4bedad2003f0cc770"} Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.363663 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5","Type":"ContainerStarted","Data":"40ff89e8515d5c5028f8867790efe34f4821c2ef396e287e61f969df42f42e86"} Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.364467 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-rhmfz" podStartSLOduration=3.396071906 podStartE2EDuration="15.364443472s" podCreationTimestamp="2025-11-27 17:38:46 +0000 UTC" firstStartedPulling="2025-11-27 17:38:48.095175138 +0000 UTC m=+1763.367632480" lastFinishedPulling="2025-11-27 17:39:00.063546694 +0000 UTC m=+1775.336004046" observedRunningTime="2025-11-27 17:39:01.357412021 +0000 UTC m=+1776.629869383" watchObservedRunningTime="2025-11-27 17:39:01.364443472 +0000 UTC m=+1776.636900824" Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.373972 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" event={"ID":"d542ca5d-c018-4f64-96f0-e515d148c51c","Type":"ContainerDied","Data":"077f2d7ab492c78999dae23c165755bb7816e0d6f334e2e11175a86a3c0aee0f"} Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.374034 4809 scope.go:117] "RemoveContainer" containerID="76cdd523a6e9bc63ffd51c4fb46910fc4a38ea7c8f838aca7f6f6ccf855f3452" Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.374185 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-w5bp6" Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.387812 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ktqz6" event={"ID":"57b62f0e-7576-4e57-9993-30000f0eee5a","Type":"ContainerStarted","Data":"7f49e17057866859ef63358c523acf92a4e1b191d6b839043accc88ec8db198c"} Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.408660 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ktqz6" podStartSLOduration=3.4545120369999998 podStartE2EDuration="15.408636095s" podCreationTimestamp="2025-11-27 17:38:46 +0000 UTC" firstStartedPulling="2025-11-27 17:38:48.161944266 +0000 UTC m=+1763.434401618" lastFinishedPulling="2025-11-27 17:39:00.116068324 +0000 UTC m=+1775.388525676" observedRunningTime="2025-11-27 17:39:01.403194387 +0000 UTC m=+1776.675651749" watchObservedRunningTime="2025-11-27 17:39:01.408636095 +0000 UTC m=+1776.681093447" Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.422103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596ec651-2088-4459-b0dd-224c67150b5d","Type":"ContainerStarted","Data":"8f46e7899255c015a731ca5d6df54e769409b95cc0ab6ec66b6c162dea49ca74"} Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.431589 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k747n" event={"ID":"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0","Type":"ContainerStarted","Data":"1cf72c60523e5210fd753ed23b3a004850d4960854894d77561575cc839a34d8"} Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.432880 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-w5bp6"] Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.441924 4809 scope.go:117] "RemoveContainer" containerID="7096d9c06d0856f926b099a322c37c172a73cb0eb76d6a9845306d90c2143807" Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.443878 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-w5bp6"] Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.461093 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-k747n" podStartSLOduration=4.461073972 podStartE2EDuration="4.461073972s" podCreationTimestamp="2025-11-27 17:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:01.448292664 +0000 UTC m=+1776.720750026" watchObservedRunningTime="2025-11-27 17:39:01.461073972 +0000 UTC m=+1776.733531324" Nov 27 17:39:01 crc kubenswrapper[4809]: I1127 17:39:01.477017 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d542ca5d-c018-4f64-96f0-e515d148c51c" path="/var/lib/kubelet/pods/d542ca5d-c018-4f64-96f0-e515d148c51c/volumes" Nov 27 17:39:02 crc kubenswrapper[4809]: I1127 17:39:02.445665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k747n" event={"ID":"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0","Type":"ContainerStarted","Data":"6b30ff49af31a2664b0b05e11743403530b80d1a8f2c1bb00688074ae847659e"} Nov 27 17:39:02 crc kubenswrapper[4809]: I1127 17:39:02.455171 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16ea53cb-7781-4900-adaa-29ac80ee734d","Type":"ContainerStarted","Data":"c0aba4c541425bf4917a6d85ccacb0dca7a5627fb03b3430ee928abc114370dc"} Nov 27 17:39:02 crc kubenswrapper[4809]: I1127 17:39:02.463060 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5","Type":"ContainerStarted","Data":"74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74"} Nov 27 17:39:02 crc kubenswrapper[4809]: I1127 17:39:02.463150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5","Type":"ContainerStarted","Data":"9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f"} Nov 27 17:39:02 crc kubenswrapper[4809]: I1127 17:39:02.463156 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerName="glance-log" containerID="cri-o://9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f" gracePeriod=30 Nov 27 17:39:02 crc kubenswrapper[4809]: I1127 17:39:02.463243 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerName="glance-httpd" containerID="cri-o://74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74" gracePeriod=30 Nov 27 17:39:02 crc kubenswrapper[4809]: I1127 17:39:02.472264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596ec651-2088-4459-b0dd-224c67150b5d","Type":"ContainerStarted","Data":"ca99702884a990489a3335cd2835a1eec23536939f25917052640d89c57c60ae"} Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.211595 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.284695 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-combined-ca-bundle\") pod \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.284767 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-logs\") pod \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.284829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.285008 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-httpd-run\") pod \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.285036 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-scripts\") pod \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.285101 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-config-data\") pod \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.285131 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgzvx\" (UniqueName: \"kubernetes.io/projected/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-kube-api-access-vgzvx\") pod \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\" (UID: \"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5\") " Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.285208 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-logs" (OuterVolumeSpecName: "logs") pod "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" (UID: "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.285492 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.286243 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" (UID: "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.291974 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-kube-api-access-vgzvx" (OuterVolumeSpecName: "kube-api-access-vgzvx") pod "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" (UID: "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5"). InnerVolumeSpecName "kube-api-access-vgzvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.292946 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-scripts" (OuterVolumeSpecName: "scripts") pod "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" (UID: "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.294863 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" (UID: "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.313507 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" (UID: "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.352918 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-config-data" (OuterVolumeSpecName: "config-data") pod "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" (UID: "86d0aafc-9af0-4eb3-aded-44a2dbefc2c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.388069 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.388103 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgzvx\" (UniqueName: \"kubernetes.io/projected/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-kube-api-access-vgzvx\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.388115 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.388152 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.388164 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.388171 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.408638 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.489841 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.492054 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16ea53cb-7781-4900-adaa-29ac80ee734d","Type":"ContainerStarted","Data":"f41f6d8e03e163ed8af381838fe5263b393903504abeadaa622f254d540f3a9e"} Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.492127 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerName="glance-log" containerID="cri-o://c0aba4c541425bf4917a6d85ccacb0dca7a5627fb03b3430ee928abc114370dc" gracePeriod=30 Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.492186 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerName="glance-httpd" containerID="cri-o://f41f6d8e03e163ed8af381838fe5263b393903504abeadaa622f254d540f3a9e" gracePeriod=30 Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.495835 4809 generic.go:334] "Generic (PLEG): container finished" podID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerID="74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74" exitCode=0 Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.495863 4809 generic.go:334] "Generic (PLEG): container finished" podID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerID="9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f" exitCode=143 Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.495866 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.495894 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5","Type":"ContainerDied","Data":"74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74"} Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.496034 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5","Type":"ContainerDied","Data":"9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f"} Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.496048 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86d0aafc-9af0-4eb3-aded-44a2dbefc2c5","Type":"ContainerDied","Data":"40ff89e8515d5c5028f8867790efe34f4821c2ef396e287e61f969df42f42e86"} Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.496066 4809 scope.go:117] "RemoveContainer" containerID="74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.535835 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.535817212 podStartE2EDuration="9.535817212s" podCreationTimestamp="2025-11-27 17:38:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:03.522809609 +0000 UTC m=+1778.795266971" watchObservedRunningTime="2025-11-27 17:39:03.535817212 +0000 UTC m=+1778.808274564" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.548904 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.552416 4809 scope.go:117] "RemoveContainer" containerID="9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.565967 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.578349 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:39:03 crc kubenswrapper[4809]: E1127 17:39:03.578781 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d542ca5d-c018-4f64-96f0-e515d148c51c" containerName="dnsmasq-dns" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.578801 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d542ca5d-c018-4f64-96f0-e515d148c51c" containerName="dnsmasq-dns" Nov 27 17:39:03 crc kubenswrapper[4809]: E1127 17:39:03.578825 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerName="glance-log" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.578832 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerName="glance-log" Nov 27 17:39:03 crc kubenswrapper[4809]: E1127 17:39:03.578845 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d542ca5d-c018-4f64-96f0-e515d148c51c" containerName="init" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.578851 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d542ca5d-c018-4f64-96f0-e515d148c51c" containerName="init" Nov 27 17:39:03 crc kubenswrapper[4809]: E1127 17:39:03.578873 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerName="glance-httpd" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.578880 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerName="glance-httpd" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.579041 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerName="glance-httpd" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.579064 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" containerName="glance-log" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.579075 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d542ca5d-c018-4f64-96f0-e515d148c51c" containerName="dnsmasq-dns" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.580180 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.583123 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.583364 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.588337 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.622375 4809 scope.go:117] "RemoveContainer" containerID="74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74" Nov 27 17:39:03 crc kubenswrapper[4809]: E1127 17:39:03.628924 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74\": container with ID starting with 74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74 not found: ID does not exist" containerID="74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.628972 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74"} err="failed to get container status \"74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74\": rpc error: code = NotFound desc = could not find container \"74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74\": container with ID starting with 74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74 not found: ID does not exist" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.628996 4809 scope.go:117] "RemoveContainer" containerID="9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f" Nov 27 17:39:03 crc kubenswrapper[4809]: E1127 17:39:03.630167 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f\": container with ID starting with 9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f not found: ID does not exist" containerID="9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.630185 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f"} err="failed to get container status \"9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f\": rpc error: code = NotFound desc = could not find container \"9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f\": container with ID starting with 9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f not found: ID does not exist" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.630200 4809 scope.go:117] "RemoveContainer" containerID="74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.630926 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74"} err="failed to get container status \"74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74\": rpc error: code = NotFound desc = could not find container \"74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74\": container with ID starting with 74f923c72fcdd33a4f05b0d5b236cef0a740dd2a0ca63fa5675a3808ffa38f74 not found: ID does not exist" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.630962 4809 scope.go:117] "RemoveContainer" containerID="9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.632306 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f"} err="failed to get container status \"9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f\": rpc error: code = NotFound desc = could not find container \"9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f\": container with ID starting with 9b73e138835c7b8c3337e4cac2205176be8c8008f70b945d9104c936d6de8c6f not found: ID does not exist" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.693935 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.695632 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.695762 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-logs\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.695788 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.695909 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.695963 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv524\" (UniqueName: \"kubernetes.io/projected/7f5d5190-5f50-4d4e-9739-250c85c0146b-kube-api-access-gv524\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.696000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.696031 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-logs\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798303 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798440 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798527 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv524\" (UniqueName: \"kubernetes.io/projected/7f5d5190-5f50-4d4e-9739-250c85c0146b-kube-api-access-gv524\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798598 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798664 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798787 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-logs\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798812 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798957 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.798817 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.799629 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.803412 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.805306 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.806165 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.815602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.819285 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv524\" (UniqueName: \"kubernetes.io/projected/7f5d5190-5f50-4d4e-9739-250c85c0146b-kube-api-access-gv524\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.835829 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:39:03 crc kubenswrapper[4809]: I1127 17:39:03.920033 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:04 crc kubenswrapper[4809]: I1127 17:39:04.508307 4809 generic.go:334] "Generic (PLEG): container finished" podID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerID="f41f6d8e03e163ed8af381838fe5263b393903504abeadaa622f254d540f3a9e" exitCode=0 Nov 27 17:39:04 crc kubenswrapper[4809]: I1127 17:39:04.508594 4809 generic.go:334] "Generic (PLEG): container finished" podID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerID="c0aba4c541425bf4917a6d85ccacb0dca7a5627fb03b3430ee928abc114370dc" exitCode=143 Nov 27 17:39:04 crc kubenswrapper[4809]: I1127 17:39:04.508390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16ea53cb-7781-4900-adaa-29ac80ee734d","Type":"ContainerDied","Data":"f41f6d8e03e163ed8af381838fe5263b393903504abeadaa622f254d540f3a9e"} Nov 27 17:39:04 crc kubenswrapper[4809]: I1127 17:39:04.508664 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16ea53cb-7781-4900-adaa-29ac80ee734d","Type":"ContainerDied","Data":"c0aba4c541425bf4917a6d85ccacb0dca7a5627fb03b3430ee928abc114370dc"} Nov 27 17:39:04 crc kubenswrapper[4809]: I1127 17:39:04.626481 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:39:04 crc kubenswrapper[4809]: W1127 17:39:04.642097 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f5d5190_5f50_4d4e_9739_250c85c0146b.slice/crio-da35c38e53fdbb5e04f7d55ebf1badf31e3dbb8a47bd115f484fd92f45321e20 WatchSource:0}: Error finding container da35c38e53fdbb5e04f7d55ebf1badf31e3dbb8a47bd115f484fd92f45321e20: Status 404 returned error can't find the container with id da35c38e53fdbb5e04f7d55ebf1badf31e3dbb8a47bd115f484fd92f45321e20 Nov 27 17:39:05 crc kubenswrapper[4809]: I1127 17:39:05.485622 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86d0aafc-9af0-4eb3-aded-44a2dbefc2c5" path="/var/lib/kubelet/pods/86d0aafc-9af0-4eb3-aded-44a2dbefc2c5/volumes" Nov 27 17:39:05 crc kubenswrapper[4809]: I1127 17:39:05.523386 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7f5d5190-5f50-4d4e-9739-250c85c0146b","Type":"ContainerStarted","Data":"da35c38e53fdbb5e04f7d55ebf1badf31e3dbb8a47bd115f484fd92f45321e20"} Nov 27 17:39:06 crc kubenswrapper[4809]: I1127 17:39:06.538706 4809 generic.go:334] "Generic (PLEG): container finished" podID="0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" containerID="6b30ff49af31a2664b0b05e11743403530b80d1a8f2c1bb00688074ae847659e" exitCode=0 Nov 27 17:39:06 crc kubenswrapper[4809]: I1127 17:39:06.538851 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k747n" event={"ID":"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0","Type":"ContainerDied","Data":"6b30ff49af31a2664b0b05e11743403530b80d1a8f2c1bb00688074ae847659e"} Nov 27 17:39:06 crc kubenswrapper[4809]: I1127 17:39:06.543392 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7f5d5190-5f50-4d4e-9739-250c85c0146b","Type":"ContainerStarted","Data":"174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045"} Nov 27 17:39:06 crc kubenswrapper[4809]: I1127 17:39:06.545961 4809 generic.go:334] "Generic (PLEG): container finished" podID="4e177c61-cc3e-444a-b6ed-aecd6d262b7c" containerID="add867ab2e890587644a3fa7893d6444487f2b07f654196452bfdcc21c669ea8" exitCode=0 Nov 27 17:39:06 crc kubenswrapper[4809]: I1127 17:39:06.545997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-rhmfz" event={"ID":"4e177c61-cc3e-444a-b6ed-aecd6d262b7c","Type":"ContainerDied","Data":"add867ab2e890587644a3fa7893d6444487f2b07f654196452bfdcc21c669ea8"} Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.109072 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.177432 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"16ea53cb-7781-4900-adaa-29ac80ee734d\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.177530 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-config-data\") pod \"16ea53cb-7781-4900-adaa-29ac80ee734d\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.177555 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-logs\") pod \"16ea53cb-7781-4900-adaa-29ac80ee734d\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.177663 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-scripts\") pod \"16ea53cb-7781-4900-adaa-29ac80ee734d\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.177834 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-httpd-run\") pod \"16ea53cb-7781-4900-adaa-29ac80ee734d\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.178024 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqjg8\" (UniqueName: \"kubernetes.io/projected/16ea53cb-7781-4900-adaa-29ac80ee734d-kube-api-access-wqjg8\") pod \"16ea53cb-7781-4900-adaa-29ac80ee734d\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.178064 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-combined-ca-bundle\") pod \"16ea53cb-7781-4900-adaa-29ac80ee734d\" (UID: \"16ea53cb-7781-4900-adaa-29ac80ee734d\") " Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.180624 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "16ea53cb-7781-4900-adaa-29ac80ee734d" (UID: "16ea53cb-7781-4900-adaa-29ac80ee734d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.180863 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-logs" (OuterVolumeSpecName: "logs") pod "16ea53cb-7781-4900-adaa-29ac80ee734d" (UID: "16ea53cb-7781-4900-adaa-29ac80ee734d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.182524 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "16ea53cb-7781-4900-adaa-29ac80ee734d" (UID: "16ea53cb-7781-4900-adaa-29ac80ee734d"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.182862 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-scripts" (OuterVolumeSpecName: "scripts") pod "16ea53cb-7781-4900-adaa-29ac80ee734d" (UID: "16ea53cb-7781-4900-adaa-29ac80ee734d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.185919 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16ea53cb-7781-4900-adaa-29ac80ee734d-kube-api-access-wqjg8" (OuterVolumeSpecName: "kube-api-access-wqjg8") pod "16ea53cb-7781-4900-adaa-29ac80ee734d" (UID: "16ea53cb-7781-4900-adaa-29ac80ee734d"). InnerVolumeSpecName "kube-api-access-wqjg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.212530 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16ea53cb-7781-4900-adaa-29ac80ee734d" (UID: "16ea53cb-7781-4900-adaa-29ac80ee734d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.231566 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-config-data" (OuterVolumeSpecName: "config-data") pod "16ea53cb-7781-4900-adaa-29ac80ee734d" (UID: "16ea53cb-7781-4900-adaa-29ac80ee734d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.279857 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.279887 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqjg8\" (UniqueName: \"kubernetes.io/projected/16ea53cb-7781-4900-adaa-29ac80ee734d-kube-api-access-wqjg8\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.279898 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.279931 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.279944 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.279953 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16ea53cb-7781-4900-adaa-29ac80ee734d-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.279960 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ea53cb-7781-4900-adaa-29ac80ee734d-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.298708 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.381513 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.560788 4809 generic.go:334] "Generic (PLEG): container finished" podID="57b62f0e-7576-4e57-9993-30000f0eee5a" containerID="7f49e17057866859ef63358c523acf92a4e1b191d6b839043accc88ec8db198c" exitCode=0 Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.560896 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ktqz6" event={"ID":"57b62f0e-7576-4e57-9993-30000f0eee5a","Type":"ContainerDied","Data":"7f49e17057866859ef63358c523acf92a4e1b191d6b839043accc88ec8db198c"} Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.564240 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16ea53cb-7781-4900-adaa-29ac80ee734d","Type":"ContainerDied","Data":"4a7efe5ca33543e6109f6bab0bbd220d4db651bdfcac38f4bedad2003f0cc770"} Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.564279 4809 scope.go:117] "RemoveContainer" containerID="f41f6d8e03e163ed8af381838fe5263b393903504abeadaa622f254d540f3a9e" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.564280 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.609264 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.617460 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.628145 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:39:07 crc kubenswrapper[4809]: E1127 17:39:07.628549 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerName="glance-log" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.628563 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerName="glance-log" Nov 27 17:39:07 crc kubenswrapper[4809]: E1127 17:39:07.628586 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerName="glance-httpd" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.628592 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerName="glance-httpd" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.628782 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerName="glance-httpd" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.628798 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ea53cb-7781-4900-adaa-29ac80ee734d" containerName="glance-log" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.629727 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.637479 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.646910 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.649629 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.687140 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-logs\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.687217 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwt68\" (UniqueName: \"kubernetes.io/projected/e345d6af-f198-42b0-9017-582c93ffeef6-kube-api-access-fwt68\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.687251 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.687283 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-scripts\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.687367 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.687384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.687422 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.687442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-config-data\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.788864 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.788916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.788974 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.789003 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-config-data\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.789032 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-logs\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.789078 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwt68\" (UniqueName: \"kubernetes.io/projected/e345d6af-f198-42b0-9017-582c93ffeef6-kube-api-access-fwt68\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.789112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.789147 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-scripts\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.790579 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-logs\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.791316 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.791359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.793838 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.794207 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-scripts\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.802003 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-config-data\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.805934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.822661 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwt68\" (UniqueName: \"kubernetes.io/projected/e345d6af-f198-42b0-9017-582c93ffeef6-kube-api-access-fwt68\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.840654 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " pod="openstack/glance-default-external-api-0" Nov 27 17:39:07 crc kubenswrapper[4809]: I1127 17:39:07.953753 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:39:09 crc kubenswrapper[4809]: I1127 17:39:09.467325 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16ea53cb-7781-4900-adaa-29ac80ee734d" path="/var/lib/kubelet/pods/16ea53cb-7781-4900-adaa-29ac80ee734d/volumes" Nov 27 17:39:12 crc kubenswrapper[4809]: I1127 17:39:12.458176 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:39:12 crc kubenswrapper[4809]: E1127 17:39:12.458843 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:39:15 crc kubenswrapper[4809]: I1127 17:39:15.637671 4809 generic.go:334] "Generic (PLEG): container finished" podID="c184a0c3-5bb0-49f0-966f-fb565fd13202" containerID="6de46e7d9a5928fd2d465277012c3423b9e297c739d698796bfebc205c6b9091" exitCode=0 Nov 27 17:39:15 crc kubenswrapper[4809]: I1127 17:39:15.637794 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-97thg" event={"ID":"c184a0c3-5bb0-49f0-966f-fb565fd13202","Type":"ContainerDied","Data":"6de46e7d9a5928fd2d465277012c3423b9e297c739d698796bfebc205c6b9091"} Nov 27 17:39:18 crc kubenswrapper[4809]: E1127 17:39:18.216563 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Nov 27 17:39:18 crc kubenswrapper[4809]: E1127 17:39:18.217068 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2vlg4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(596ec651-2088-4459-b0dd-224c67150b5d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.295900 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k747n" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.304976 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.317002 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-rhmfz" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.386936 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spbpw\" (UniqueName: \"kubernetes.io/projected/57b62f0e-7576-4e57-9993-30000f0eee5a-kube-api-access-spbpw\") pod \"57b62f0e-7576-4e57-9993-30000f0eee5a\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387028 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-combined-ca-bundle\") pod \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387062 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-config-data\") pod \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387126 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-scripts\") pod \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387154 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-credential-keys\") pod \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387179 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-combined-ca-bundle\") pod \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387416 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-config-data\") pod \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387442 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-db-sync-config-data\") pod \"57b62f0e-7576-4e57-9993-30000f0eee5a\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387467 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwdz6\" (UniqueName: \"kubernetes.io/projected/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-kube-api-access-vwdz6\") pod \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387506 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-combined-ca-bundle\") pod \"57b62f0e-7576-4e57-9993-30000f0eee5a\" (UID: \"57b62f0e-7576-4e57-9993-30000f0eee5a\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387566 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-logs\") pod \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\" (UID: \"4e177c61-cc3e-444a-b6ed-aecd6d262b7c\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-fernet-keys\") pod \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387653 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-scripts\") pod \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387770 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2fbt\" (UniqueName: \"kubernetes.io/projected/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-kube-api-access-l2fbt\") pod \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\" (UID: \"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0\") " Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.387973 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-logs" (OuterVolumeSpecName: "logs") pod "4e177c61-cc3e-444a-b6ed-aecd6d262b7c" (UID: "4e177c61-cc3e-444a-b6ed-aecd6d262b7c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.388538 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.393169 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-scripts" (OuterVolumeSpecName: "scripts") pod "4e177c61-cc3e-444a-b6ed-aecd6d262b7c" (UID: "4e177c61-cc3e-444a-b6ed-aecd6d262b7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.393326 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" (UID: "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.393736 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b62f0e-7576-4e57-9993-30000f0eee5a-kube-api-access-spbpw" (OuterVolumeSpecName: "kube-api-access-spbpw") pod "57b62f0e-7576-4e57-9993-30000f0eee5a" (UID: "57b62f0e-7576-4e57-9993-30000f0eee5a"). InnerVolumeSpecName "kube-api-access-spbpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.395375 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "57b62f0e-7576-4e57-9993-30000f0eee5a" (UID: "57b62f0e-7576-4e57-9993-30000f0eee5a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.396073 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-kube-api-access-l2fbt" (OuterVolumeSpecName: "kube-api-access-l2fbt") pod "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" (UID: "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0"). InnerVolumeSpecName "kube-api-access-l2fbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.396379 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" (UID: "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.403956 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-kube-api-access-vwdz6" (OuterVolumeSpecName: "kube-api-access-vwdz6") pod "4e177c61-cc3e-444a-b6ed-aecd6d262b7c" (UID: "4e177c61-cc3e-444a-b6ed-aecd6d262b7c"). InnerVolumeSpecName "kube-api-access-vwdz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.415327 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-config-data" (OuterVolumeSpecName: "config-data") pod "4e177c61-cc3e-444a-b6ed-aecd6d262b7c" (UID: "4e177c61-cc3e-444a-b6ed-aecd6d262b7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.417029 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57b62f0e-7576-4e57-9993-30000f0eee5a" (UID: "57b62f0e-7576-4e57-9993-30000f0eee5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.418868 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e177c61-cc3e-444a-b6ed-aecd6d262b7c" (UID: "4e177c61-cc3e-444a-b6ed-aecd6d262b7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.412523 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-scripts" (OuterVolumeSpecName: "scripts") pod "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" (UID: "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.424663 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-config-data" (OuterVolumeSpecName: "config-data") pod "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" (UID: "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.428060 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" (UID: "0bbb5fa1-67a7-4d6e-9354-36598e97d9d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491070 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491109 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491123 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491135 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491488 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491506 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491518 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491530 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwdz6\" (UniqueName: \"kubernetes.io/projected/4e177c61-cc3e-444a-b6ed-aecd6d262b7c-kube-api-access-vwdz6\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491545 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b62f0e-7576-4e57-9993-30000f0eee5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491752 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491788 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491802 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2fbt\" (UniqueName: \"kubernetes.io/projected/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0-kube-api-access-l2fbt\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.491818 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spbpw\" (UniqueName: \"kubernetes.io/projected/57b62f0e-7576-4e57-9993-30000f0eee5a-kube-api-access-spbpw\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.663316 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ktqz6" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.663319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ktqz6" event={"ID":"57b62f0e-7576-4e57-9993-30000f0eee5a","Type":"ContainerDied","Data":"aa8c7ebb5170f3af85b5c2194ff1184ca2726e75149b6dd707629693f84d927d"} Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.663369 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa8c7ebb5170f3af85b5c2194ff1184ca2726e75149b6dd707629693f84d927d" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.665936 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k747n" event={"ID":"0bbb5fa1-67a7-4d6e-9354-36598e97d9d0","Type":"ContainerDied","Data":"1cf72c60523e5210fd753ed23b3a004850d4960854894d77561575cc839a34d8"} Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.665975 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cf72c60523e5210fd753ed23b3a004850d4960854894d77561575cc839a34d8" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.666029 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k747n" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.674290 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-rhmfz" event={"ID":"4e177c61-cc3e-444a-b6ed-aecd6d262b7c","Type":"ContainerDied","Data":"cf04a7d3faa3280057a00eab61b82df001134464cb7cf98bec14a4048794a7bc"} Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.674335 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf04a7d3faa3280057a00eab61b82df001134464cb7cf98bec14a4048794a7bc" Nov 27 17:39:18 crc kubenswrapper[4809]: I1127 17:39:18.674311 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-rhmfz" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.235434 4809 scope.go:117] "RemoveContainer" containerID="c0aba4c541425bf4917a6d85ccacb0dca7a5627fb03b3430ee928abc114370dc" Nov 27 17:39:19 crc kubenswrapper[4809]: E1127 17:39:19.274568 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 27 17:39:19 crc kubenswrapper[4809]: E1127 17:39:19.274854 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hrv44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-t2gnx_openstack(9a3b3964-4402-49a4-86ba-e6e7909082ec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 17:39:19 crc kubenswrapper[4809]: E1127 17:39:19.276076 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-t2gnx" podUID="9a3b3964-4402-49a4-86ba-e6e7909082ec" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.420015 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-97thg" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.432919 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-856cbdcf9c-lws7x"] Nov 27 17:39:19 crc kubenswrapper[4809]: E1127 17:39:19.433972 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c184a0c3-5bb0-49f0-966f-fb565fd13202" containerName="neutron-db-sync" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.435977 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c184a0c3-5bb0-49f0-966f-fb565fd13202" containerName="neutron-db-sync" Nov 27 17:39:19 crc kubenswrapper[4809]: E1127 17:39:19.436127 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b62f0e-7576-4e57-9993-30000f0eee5a" containerName="barbican-db-sync" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.436369 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b62f0e-7576-4e57-9993-30000f0eee5a" containerName="barbican-db-sync" Nov 27 17:39:19 crc kubenswrapper[4809]: E1127 17:39:19.448463 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" containerName="keystone-bootstrap" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.448500 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" containerName="keystone-bootstrap" Nov 27 17:39:19 crc kubenswrapper[4809]: E1127 17:39:19.448572 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e177c61-cc3e-444a-b6ed-aecd6d262b7c" containerName="placement-db-sync" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.448583 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e177c61-cc3e-444a-b6ed-aecd6d262b7c" containerName="placement-db-sync" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.449520 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b62f0e-7576-4e57-9993-30000f0eee5a" containerName="barbican-db-sync" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.449557 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" containerName="keystone-bootstrap" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.449582 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c184a0c3-5bb0-49f0-966f-fb565fd13202" containerName="neutron-db-sync" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.449612 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e177c61-cc3e-444a-b6ed-aecd6d262b7c" containerName="placement-db-sync" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.450722 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.457900 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-l4mwz" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.458773 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.459032 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.459293 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.459518 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.459702 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.533911 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-combined-ca-bundle\") pod \"c184a0c3-5bb0-49f0-966f-fb565fd13202\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.533965 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kk75\" (UniqueName: \"kubernetes.io/projected/c184a0c3-5bb0-49f0-966f-fb565fd13202-kube-api-access-5kk75\") pod \"c184a0c3-5bb0-49f0-966f-fb565fd13202\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.533988 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-config\") pod \"c184a0c3-5bb0-49f0-966f-fb565fd13202\" (UID: \"c184a0c3-5bb0-49f0-966f-fb565fd13202\") " Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.534344 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-combined-ca-bundle\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.535313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-internal-tls-certs\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.535605 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-public-tls-certs\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.535732 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-credential-keys\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.535891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hpkp\" (UniqueName: \"kubernetes.io/projected/5ddd7431-85e1-4f2e-ae95-f00162cbe120-kube-api-access-6hpkp\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.535941 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-scripts\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.535963 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-fernet-keys\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.536016 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-config-data\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.538027 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-856cbdcf9c-lws7x"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.560287 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c184a0c3-5bb0-49f0-966f-fb565fd13202-kube-api-access-5kk75" (OuterVolumeSpecName: "kube-api-access-5kk75") pod "c184a0c3-5bb0-49f0-966f-fb565fd13202" (UID: "c184a0c3-5bb0-49f0-966f-fb565fd13202"). InnerVolumeSpecName "kube-api-access-5kk75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.605507 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-556f88f66d-dk28p"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.607520 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.624181 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-556f88f66d-dk28p"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.625663 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.625931 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-lr45g" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.626173 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.626275 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.626478 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.641828 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-65b4758f4c-5krfp"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.645473 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656001 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656380 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-84d69cc6b6-vqlb7"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656441 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af7083a1-8a26-49b0-8b7a-7b412592ed60-config-data-custom\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656494 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7083a1-8a26-49b0-8b7a-7b412592ed60-combined-ca-bundle\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656518 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af7083a1-8a26-49b0-8b7a-7b412592ed60-logs\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656563 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-public-tls-certs\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656610 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-credential-keys\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656648 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hpkp\" (UniqueName: \"kubernetes.io/projected/5ddd7431-85e1-4f2e-ae95-f00162cbe120-kube-api-access-6hpkp\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656717 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-scripts\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656755 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-fernet-keys\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656776 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdqqn\" (UniqueName: \"kubernetes.io/projected/af7083a1-8a26-49b0-8b7a-7b412592ed60-kube-api-access-wdqqn\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-config-data\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656847 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-combined-ca-bundle\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656885 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7083a1-8a26-49b0-8b7a-7b412592ed60-config-data\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-internal-tls-certs\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.656983 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kk75\" (UniqueName: \"kubernetes.io/projected/c184a0c3-5bb0-49f0-966f-fb565fd13202-kube-api-access-5kk75\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.657970 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.663251 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.663697 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.665312 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c184a0c3-5bb0-49f0-966f-fb565fd13202" (UID: "c184a0c3-5bb0-49f0-966f-fb565fd13202"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.667560 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-combined-ca-bundle\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.668392 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-public-tls-certs\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.671567 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-scripts\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.672558 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-config-data\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.672727 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-config" (OuterVolumeSpecName: "config") pod "c184a0c3-5bb0-49f0-966f-fb565fd13202" (UID: "c184a0c3-5bb0-49f0-966f-fb565fd13202"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.674413 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qsvkg" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.690360 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hpkp\" (UniqueName: \"kubernetes.io/projected/5ddd7431-85e1-4f2e-ae95-f00162cbe120-kube-api-access-6hpkp\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.693464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-fernet-keys\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.696936 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-internal-tls-certs\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.702314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-97thg" event={"ID":"c184a0c3-5bb0-49f0-966f-fb565fd13202","Type":"ContainerDied","Data":"d074b23df3560cd4234217ac9340a476f2216b05af00993aa1a9311608f6d591"} Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.702352 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d074b23df3560cd4234217ac9340a476f2216b05af00993aa1a9311608f6d591" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.702419 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-97thg" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.706086 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-65b4758f4c-5krfp"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.706573 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5ddd7431-85e1-4f2e-ae95-f00162cbe120-credential-keys\") pod \"keystone-856cbdcf9c-lws7x\" (UID: \"5ddd7431-85e1-4f2e-ae95-f00162cbe120\") " pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.734345 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-84d69cc6b6-vqlb7"] Nov 27 17:39:19 crc kubenswrapper[4809]: E1127 17:39:19.734465 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-t2gnx" podUID="9a3b3964-4402-49a4-86ba-e6e7909082ec" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.762637 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbw8k\" (UniqueName: \"kubernetes.io/projected/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-kube-api-access-zbw8k\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.762692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-internal-tls-certs\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.762711 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-combined-ca-bundle\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.762836 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs5p2\" (UniqueName: \"kubernetes.io/projected/a43816b3-bfa7-4570-b082-9b217d660416-kube-api-access-fs5p2\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.762914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-config-data\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.762933 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43816b3-bfa7-4570-b082-9b217d660416-logs\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.762976 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-config-data\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.763001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7083a1-8a26-49b0-8b7a-7b412592ed60-config-data\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.763023 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-public-tls-certs\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.763073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af7083a1-8a26-49b0-8b7a-7b412592ed60-config-data-custom\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.763096 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-combined-ca-bundle\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.763126 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7083a1-8a26-49b0-8b7a-7b412592ed60-combined-ca-bundle\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.763152 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af7083a1-8a26-49b0-8b7a-7b412592ed60-logs\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.763287 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-scripts\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.763324 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-logs\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.764333 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-config-data-custom\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.764388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdqqn\" (UniqueName: \"kubernetes.io/projected/af7083a1-8a26-49b0-8b7a-7b412592ed60-kube-api-access-wdqqn\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.764470 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.764481 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c184a0c3-5bb0-49f0-966f-fb565fd13202-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.766538 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af7083a1-8a26-49b0-8b7a-7b412592ed60-logs\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.772865 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7083a1-8a26-49b0-8b7a-7b412592ed60-combined-ca-bundle\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.773501 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af7083a1-8a26-49b0-8b7a-7b412592ed60-config-data-custom\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.783494 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7083a1-8a26-49b0-8b7a-7b412592ed60-config-data\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.789598 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdqqn\" (UniqueName: \"kubernetes.io/projected/af7083a1-8a26-49b0-8b7a-7b412592ed60-kube-api-access-wdqqn\") pod \"barbican-worker-65b4758f4c-5krfp\" (UID: \"af7083a1-8a26-49b0-8b7a-7b412592ed60\") " pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.820251 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-qpdtc"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.824782 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.832077 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.835784 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-qpdtc"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.865828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-combined-ca-bundle\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.865890 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-scripts\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.865917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-logs\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866119 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-config-data-custom\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866252 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbw8k\" (UniqueName: \"kubernetes.io/projected/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-kube-api-access-zbw8k\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866313 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-internal-tls-certs\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866337 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-combined-ca-bundle\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866400 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs5p2\" (UniqueName: \"kubernetes.io/projected/a43816b3-bfa7-4570-b082-9b217d660416-kube-api-access-fs5p2\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866608 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-config-data\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866637 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43816b3-bfa7-4570-b082-9b217d660416-logs\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866755 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-config-data\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-public-tls-certs\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.866949 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-logs\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.870281 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43816b3-bfa7-4570-b082-9b217d660416-logs\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.870685 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6688bc68c4-s95bj"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.870878 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-scripts\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.873150 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-combined-ca-bundle\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.875036 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-combined-ca-bundle\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.875324 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-internal-tls-certs\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.876840 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-public-tls-certs\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.877423 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.880349 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43816b3-bfa7-4570-b082-9b217d660416-config-data\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.882153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-config-data-custom\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.883035 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6688bc68c4-s95bj"] Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.884332 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.884852 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-config-data\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.893203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbw8k\" (UniqueName: \"kubernetes.io/projected/9bc3b7a8-110d-481e-95d5-3d9034c0b2f9-kube-api-access-zbw8k\") pod \"barbican-keystone-listener-84d69cc6b6-vqlb7\" (UID: \"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9\") " pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.894578 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs5p2\" (UniqueName: \"kubernetes.io/projected/a43816b3-bfa7-4570-b082-9b217d660416-kube-api-access-fs5p2\") pod \"placement-556f88f66d-dk28p\" (UID: \"a43816b3-bfa7-4570-b082-9b217d660416\") " pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.975477 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-config\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.975537 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c406df20-1687-4079-8b4b-8466a334e9a8-logs\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.975560 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lvdj\" (UniqueName: \"kubernetes.io/projected/c406df20-1687-4079-8b4b-8466a334e9a8-kube-api-access-2lvdj\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.975692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.975781 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.975813 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data-custom\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.975881 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-combined-ca-bundle\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.976031 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.976105 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.976237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:19 crc kubenswrapper[4809]: I1127 17:39:19.976612 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hkl2\" (UniqueName: \"kubernetes.io/projected/f8f94287-4435-47b7-a903-bebb76b6dfe5-kube-api-access-8hkl2\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.002032 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.019209 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.056968 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-65b4758f4c-5krfp" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.072689 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.079899 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080019 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080174 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hkl2\" (UniqueName: \"kubernetes.io/projected/f8f94287-4435-47b7-a903-bebb76b6dfe5-kube-api-access-8hkl2\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080480 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-config\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lvdj\" (UniqueName: \"kubernetes.io/projected/c406df20-1687-4079-8b4b-8466a334e9a8-kube-api-access-2lvdj\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c406df20-1687-4079-8b4b-8466a334e9a8-logs\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080566 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080595 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080638 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data-custom\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.080682 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-combined-ca-bundle\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.082274 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-config\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.084705 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c406df20-1687-4079-8b4b-8466a334e9a8-logs\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.085266 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.085273 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.086249 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-combined-ca-bundle\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.086553 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.086664 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.091008 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data-custom\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.098785 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.099140 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lvdj\" (UniqueName: \"kubernetes.io/projected/c406df20-1687-4079-8b4b-8466a334e9a8-kube-api-access-2lvdj\") pod \"barbican-api-6688bc68c4-s95bj\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.104676 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hkl2\" (UniqueName: \"kubernetes.io/projected/f8f94287-4435-47b7-a903-bebb76b6dfe5-kube-api-access-8hkl2\") pod \"dnsmasq-dns-59d5ff467f-qpdtc\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.151651 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.225487 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.399960 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-856cbdcf9c-lws7x"] Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.646795 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-556f88f66d-dk28p"] Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.669538 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-qpdtc"] Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.701213 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-qslw2"] Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.704934 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.726537 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-qslw2"] Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.816151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7f5d5190-5f50-4d4e-9739-250c85c0146b","Type":"ContainerStarted","Data":"fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97"} Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.816827 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-config\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.816968 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.817099 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.817158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.817237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.817429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h72w4\" (UniqueName: \"kubernetes.io/projected/b1890991-c957-47ca-89d4-307e357808d4-kube-api-access-h72w4\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.828415 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-856cbdcf9c-lws7x" event={"ID":"5ddd7431-85e1-4f2e-ae95-f00162cbe120","Type":"ContainerStarted","Data":"2259424d605c7b28c12a14aecec1b9414368c938fbeb91995cd210f225101008"} Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.831069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e345d6af-f198-42b0-9017-582c93ffeef6","Type":"ContainerStarted","Data":"2d9013182302b4f4ae1bb9bdc3942283fe488e5430143ea4669a08d9de4f7b3d"} Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.833168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-556f88f66d-dk28p" event={"ID":"a43816b3-bfa7-4570-b082-9b217d660416","Type":"ContainerStarted","Data":"802cf3b61e54e98b15f4597d9e458e1d0605e5e8b52f021ac3d4493f975c9e88"} Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.872241 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cffcf7bfb-7p9tr"] Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.879849 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.891978 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.892222 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.892761 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.895453 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cffcf7bfb-7p9tr"] Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.896050 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wngmx" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.928857 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=17.928833807 podStartE2EDuration="17.928833807s" podCreationTimestamp="2025-11-27 17:39:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:20.844599235 +0000 UTC m=+1796.117056587" watchObservedRunningTime="2025-11-27 17:39:20.928833807 +0000 UTC m=+1796.201291159" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929014 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929062 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929084 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-httpd-config\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929140 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnts9\" (UniqueName: \"kubernetes.io/projected/013eaaea-f9de-4e5c-9647-a5df367b7709-kube-api-access-cnts9\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929409 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h72w4\" (UniqueName: \"kubernetes.io/projected/b1890991-c957-47ca-89d4-307e357808d4-kube-api-access-h72w4\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929466 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-ovndb-tls-certs\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-config\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929560 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929613 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-config\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.929647 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-combined-ca-bundle\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.931883 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.932804 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.933291 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-config\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.936630 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.947559 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.960791 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-qpdtc"] Nov 27 17:39:20 crc kubenswrapper[4809]: I1127 17:39:20.965991 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h72w4\" (UniqueName: \"kubernetes.io/projected/b1890991-c957-47ca-89d4-307e357808d4-kube-api-access-h72w4\") pod \"dnsmasq-dns-75c8ddd69c-qslw2\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.018185 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-65b4758f4c-5krfp"] Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.030882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnts9\" (UniqueName: \"kubernetes.io/projected/013eaaea-f9de-4e5c-9647-a5df367b7709-kube-api-access-cnts9\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.030998 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-ovndb-tls-certs\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.031070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-config\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.031096 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-combined-ca-bundle\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.031137 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-httpd-config\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.037845 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-config\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.039317 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-combined-ca-bundle\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.039421 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-ovndb-tls-certs\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.049261 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-httpd-config\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.086589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnts9\" (UniqueName: \"kubernetes.io/projected/013eaaea-f9de-4e5c-9647-a5df367b7709-kube-api-access-cnts9\") pod \"neutron-6cffcf7bfb-7p9tr\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.121408 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-84d69cc6b6-vqlb7"] Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.166973 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.218534 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.251809 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6688bc68c4-s95bj"] Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.855150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-856cbdcf9c-lws7x" event={"ID":"5ddd7431-85e1-4f2e-ae95-f00162cbe120","Type":"ContainerStarted","Data":"d813e7c0326ccd55c308733a6886c6b0d1cfb069f99ba3920294b87cc99efa00"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.855319 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.879453 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-856cbdcf9c-lws7x" podStartSLOduration=2.87943352 podStartE2EDuration="2.87943352s" podCreationTimestamp="2025-11-27 17:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:21.876597033 +0000 UTC m=+1797.149054385" watchObservedRunningTime="2025-11-27 17:39:21.87943352 +0000 UTC m=+1797.151890872" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.881830 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e345d6af-f198-42b0-9017-582c93ffeef6","Type":"ContainerStarted","Data":"4c303b3aa29b3f559a9eedeb7784e5a607de6c46cda79f3c034b0422b71cbd76"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.891153 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-556f88f66d-dk28p" event={"ID":"a43816b3-bfa7-4570-b082-9b217d660416","Type":"ContainerStarted","Data":"9ab3829ad3d3cc6df3bef5d6d1ef5c7a61cf0ecbb1cde97719e4f357f70b5f9e"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.891212 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-556f88f66d-dk28p" event={"ID":"a43816b3-bfa7-4570-b082-9b217d660416","Type":"ContainerStarted","Data":"2f702254785b7c2fcf8b353766616f8f5481fa8922e934aecbad7c5513cd1d4b"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.892497 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.892527 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.895086 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-65b4758f4c-5krfp" event={"ID":"af7083a1-8a26-49b0-8b7a-7b412592ed60","Type":"ContainerStarted","Data":"e92be2021ac7afc27e708abb308afdea55a17428cd2ffdc3805df83f3d504bae"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.905156 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6688bc68c4-s95bj" event={"ID":"c406df20-1687-4079-8b4b-8466a334e9a8","Type":"ContainerStarted","Data":"13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.905198 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6688bc68c4-s95bj" event={"ID":"c406df20-1687-4079-8b4b-8466a334e9a8","Type":"ContainerStarted","Data":"456cef95bf8eddad75cc79095d63d87fca5c1478dd0cd8694b58838668737057"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.909879 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" event={"ID":"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9","Type":"ContainerStarted","Data":"5768fc6fde773ef5eb1e7eb3507fbe4cd6d228118009d6004cbfbd6aac01da9f"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.912578 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-556f88f66d-dk28p" podStartSLOduration=2.912560642 podStartE2EDuration="2.912560642s" podCreationTimestamp="2025-11-27 17:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:21.91135217 +0000 UTC m=+1797.183809522" watchObservedRunningTime="2025-11-27 17:39:21.912560642 +0000 UTC m=+1797.185018004" Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.915543 4809 generic.go:334] "Generic (PLEG): container finished" podID="f8f94287-4435-47b7-a903-bebb76b6dfe5" containerID="cc15b52c5d2baf7118833c75e2bc429d20036108603797aabf4b43626c8dd498" exitCode=0 Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.917075 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" event={"ID":"f8f94287-4435-47b7-a903-bebb76b6dfe5","Type":"ContainerDied","Data":"cc15b52c5d2baf7118833c75e2bc429d20036108603797aabf4b43626c8dd498"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.917118 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" event={"ID":"f8f94287-4435-47b7-a903-bebb76b6dfe5","Type":"ContainerStarted","Data":"d6ac4e452482d07928e9d25331220a83631c55a11579cba8dca395723a509003"} Nov 27 17:39:21 crc kubenswrapper[4809]: I1127 17:39:21.968934 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-qslw2"] Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.178772 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cffcf7bfb-7p9tr"] Nov 27 17:39:22 crc kubenswrapper[4809]: W1127 17:39:22.209966 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013eaaea_f9de_4e5c_9647_a5df367b7709.slice/crio-798ad1c59a1972200e09171373a8932a578d98f2bfd8415b5e8ef0b97bddb5ae WatchSource:0}: Error finding container 798ad1c59a1972200e09171373a8932a578d98f2bfd8415b5e8ef0b97bddb5ae: Status 404 returned error can't find the container with id 798ad1c59a1972200e09171373a8932a578d98f2bfd8415b5e8ef0b97bddb5ae Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.571881 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.672659 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-svc\") pod \"f8f94287-4435-47b7-a903-bebb76b6dfe5\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.672748 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-nb\") pod \"f8f94287-4435-47b7-a903-bebb76b6dfe5\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.672773 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-sb\") pod \"f8f94287-4435-47b7-a903-bebb76b6dfe5\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.672939 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hkl2\" (UniqueName: \"kubernetes.io/projected/f8f94287-4435-47b7-a903-bebb76b6dfe5-kube-api-access-8hkl2\") pod \"f8f94287-4435-47b7-a903-bebb76b6dfe5\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.672983 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-config\") pod \"f8f94287-4435-47b7-a903-bebb76b6dfe5\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.673017 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-swift-storage-0\") pod \"f8f94287-4435-47b7-a903-bebb76b6dfe5\" (UID: \"f8f94287-4435-47b7-a903-bebb76b6dfe5\") " Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.720073 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-config" (OuterVolumeSpecName: "config") pod "f8f94287-4435-47b7-a903-bebb76b6dfe5" (UID: "f8f94287-4435-47b7-a903-bebb76b6dfe5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.720721 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f8f94287-4435-47b7-a903-bebb76b6dfe5" (UID: "f8f94287-4435-47b7-a903-bebb76b6dfe5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.722982 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f8f94287-4435-47b7-a903-bebb76b6dfe5" (UID: "f8f94287-4435-47b7-a903-bebb76b6dfe5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.725937 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f8f94287-4435-47b7-a903-bebb76b6dfe5" (UID: "f8f94287-4435-47b7-a903-bebb76b6dfe5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.729644 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f8f94287-4435-47b7-a903-bebb76b6dfe5" (UID: "f8f94287-4435-47b7-a903-bebb76b6dfe5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.731165 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f94287-4435-47b7-a903-bebb76b6dfe5-kube-api-access-8hkl2" (OuterVolumeSpecName: "kube-api-access-8hkl2") pod "f8f94287-4435-47b7-a903-bebb76b6dfe5" (UID: "f8f94287-4435-47b7-a903-bebb76b6dfe5"). InnerVolumeSpecName "kube-api-access-8hkl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.775337 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hkl2\" (UniqueName: \"kubernetes.io/projected/f8f94287-4435-47b7-a903-bebb76b6dfe5-kube-api-access-8hkl2\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.775375 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.775386 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.775395 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.775405 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.775412 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8f94287-4435-47b7-a903-bebb76b6dfe5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.928580 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" event={"ID":"f8f94287-4435-47b7-a903-bebb76b6dfe5","Type":"ContainerDied","Data":"d6ac4e452482d07928e9d25331220a83631c55a11579cba8dca395723a509003"} Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.928600 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-qpdtc" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.928640 4809 scope.go:117] "RemoveContainer" containerID="cc15b52c5d2baf7118833c75e2bc429d20036108603797aabf4b43626c8dd498" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.931037 4809 generic.go:334] "Generic (PLEG): container finished" podID="b1890991-c957-47ca-89d4-307e357808d4" containerID="a452e9686dda99a1da061cb064f877094c9ade39be9a185d42144a0a98233917" exitCode=0 Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.931110 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" event={"ID":"b1890991-c957-47ca-89d4-307e357808d4","Type":"ContainerDied","Data":"a452e9686dda99a1da061cb064f877094c9ade39be9a185d42144a0a98233917"} Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.931135 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" event={"ID":"b1890991-c957-47ca-89d4-307e357808d4","Type":"ContainerStarted","Data":"573381ea3f7970cae0be63e0c7bcd4ef310d777b25754e47af3aecd4d1a71ec2"} Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.941139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e345d6af-f198-42b0-9017-582c93ffeef6","Type":"ContainerStarted","Data":"6bd864cab97711e66dabfdb917ca4f1b88372afed903288f05688afb3b95b68f"} Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.946676 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6688bc68c4-s95bj" event={"ID":"c406df20-1687-4079-8b4b-8466a334e9a8","Type":"ContainerStarted","Data":"4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9"} Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.946799 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.946956 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.950885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffcf7bfb-7p9tr" event={"ID":"013eaaea-f9de-4e5c-9647-a5df367b7709","Type":"ContainerStarted","Data":"9035331522b1447565598f7473b7ab808d2ec9ca68cd09992c867d8b5192020b"} Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.950922 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffcf7bfb-7p9tr" event={"ID":"013eaaea-f9de-4e5c-9647-a5df367b7709","Type":"ContainerStarted","Data":"798ad1c59a1972200e09171373a8932a578d98f2bfd8415b5e8ef0b97bddb5ae"} Nov 27 17:39:22 crc kubenswrapper[4809]: I1127 17:39:22.982492 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6688bc68c4-s95bj" podStartSLOduration=3.982469523 podStartE2EDuration="3.982469523s" podCreationTimestamp="2025-11-27 17:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:22.97685812 +0000 UTC m=+1798.249315472" watchObservedRunningTime="2025-11-27 17:39:22.982469523 +0000 UTC m=+1798.254926875" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.021339 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-qpdtc"] Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.029780 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-qpdtc"] Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.043809 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=16.043784252 podStartE2EDuration="16.043784252s" podCreationTimestamp="2025-11-27 17:39:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:23.04041015 +0000 UTC m=+1798.312867502" watchObservedRunningTime="2025-11-27 17:39:23.043784252 +0000 UTC m=+1798.316241604" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.474188 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8f94287-4435-47b7-a903-bebb76b6dfe5" path="/var/lib/kubelet/pods/f8f94287-4435-47b7-a903-bebb76b6dfe5/volumes" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.615513 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-65c6b6d897-rhmzw"] Nov 27 17:39:23 crc kubenswrapper[4809]: E1127 17:39:23.615981 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f94287-4435-47b7-a903-bebb76b6dfe5" containerName="init" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.616004 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f94287-4435-47b7-a903-bebb76b6dfe5" containerName="init" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.616266 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f94287-4435-47b7-a903-bebb76b6dfe5" containerName="init" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.617460 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.620232 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.620245 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.642567 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65c6b6d897-rhmzw"] Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.704238 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-combined-ca-bundle\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.704298 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-config\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.704331 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-httpd-config\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.704355 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-ovndb-tls-certs\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.704512 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-public-tls-certs\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.704636 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s74b2\" (UniqueName: \"kubernetes.io/projected/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-kube-api-access-s74b2\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.704687 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-internal-tls-certs\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.806517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-combined-ca-bundle\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.806598 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-config\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.806637 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-httpd-config\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.806662 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-ovndb-tls-certs\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.806701 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-public-tls-certs\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.806769 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s74b2\" (UniqueName: \"kubernetes.io/projected/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-kube-api-access-s74b2\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.806918 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-internal-tls-certs\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.812840 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-ovndb-tls-certs\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.812904 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-config\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.813370 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-combined-ca-bundle\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.813687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-httpd-config\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.822790 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-public-tls-certs\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.824073 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-internal-tls-certs\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.831078 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s74b2\" (UniqueName: \"kubernetes.io/projected/1a7bab33-46e8-4e18-a0c2-1ca25e22493f-kube-api-access-s74b2\") pod \"neutron-65c6b6d897-rhmzw\" (UID: \"1a7bab33-46e8-4e18-a0c2-1ca25e22493f\") " pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.920789 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.920842 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.938900 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.954946 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.964158 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffcf7bfb-7p9tr" event={"ID":"013eaaea-f9de-4e5c-9647-a5df367b7709","Type":"ContainerStarted","Data":"51d28b59af8a554b584b8f720373c960b56ded2c93ae0456729f8b49d97db81c"} Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.965032 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.968471 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" event={"ID":"b1890991-c957-47ca-89d4-307e357808d4","Type":"ContainerStarted","Data":"4740da13b9f9a3dab0f9167c31fb3d7f2507674dae05217ffc335c739389f02e"} Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.968918 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:23 crc kubenswrapper[4809]: I1127 17:39:23.978338 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:24 crc kubenswrapper[4809]: I1127 17:39:24.009911 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cffcf7bfb-7p9tr" podStartSLOduration=4.009886498 podStartE2EDuration="4.009886498s" podCreationTimestamp="2025-11-27 17:39:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:23.991585019 +0000 UTC m=+1799.264042371" watchObservedRunningTime="2025-11-27 17:39:24.009886498 +0000 UTC m=+1799.282343860" Nov 27 17:39:24 crc kubenswrapper[4809]: I1127 17:39:24.020185 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" podStartSLOduration=4.020170587 podStartE2EDuration="4.020170587s" podCreationTimestamp="2025-11-27 17:39:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:24.011553202 +0000 UTC m=+1799.284010554" watchObservedRunningTime="2025-11-27 17:39:24.020170587 +0000 UTC m=+1799.292627939" Nov 27 17:39:24 crc kubenswrapper[4809]: I1127 17:39:24.976967 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:24 crc kubenswrapper[4809]: I1127 17:39:24.977005 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:25 crc kubenswrapper[4809]: I1127 17:39:25.457520 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:39:26 crc kubenswrapper[4809]: E1127 17:39:25.458194 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.076337 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5bfc74dc5d-snbcn"] Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.078153 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.080138 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.080397 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.089911 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5bfc74dc5d-snbcn"] Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.151257 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-public-tls-certs\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.151304 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b58l\" (UniqueName: \"kubernetes.io/projected/fea753ab-13fb-49aa-a430-6c5db50f1e6b-kube-api-access-7b58l\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.151399 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fea753ab-13fb-49aa-a430-6c5db50f1e6b-logs\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.151444 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-internal-tls-certs\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.151468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-config-data-custom\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.151572 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-combined-ca-bundle\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.151881 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-config-data\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.254021 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-config-data\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.254091 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-public-tls-certs\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.254135 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b58l\" (UniqueName: \"kubernetes.io/projected/fea753ab-13fb-49aa-a430-6c5db50f1e6b-kube-api-access-7b58l\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.254222 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fea753ab-13fb-49aa-a430-6c5db50f1e6b-logs\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.254714 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fea753ab-13fb-49aa-a430-6c5db50f1e6b-logs\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.254278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-internal-tls-certs\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.254834 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-config-data-custom\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.254860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-combined-ca-bundle\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.267034 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-combined-ca-bundle\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.267034 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-internal-tls-certs\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.267604 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-public-tls-certs\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.268207 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-config-data-custom\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.268435 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea753ab-13fb-49aa-a430-6c5db50f1e6b-config-data\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.273964 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b58l\" (UniqueName: \"kubernetes.io/projected/fea753ab-13fb-49aa-a430-6c5db50f1e6b-kube-api-access-7b58l\") pod \"barbican-api-5bfc74dc5d-snbcn\" (UID: \"fea753ab-13fb-49aa-a430-6c5db50f1e6b\") " pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:26 crc kubenswrapper[4809]: I1127 17:39:26.393843 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:27 crc kubenswrapper[4809]: I1127 17:39:27.437036 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:27 crc kubenswrapper[4809]: I1127 17:39:27.437146 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 17:39:27 crc kubenswrapper[4809]: I1127 17:39:27.438665 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 27 17:39:27 crc kubenswrapper[4809]: I1127 17:39:27.955005 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 27 17:39:27 crc kubenswrapper[4809]: I1127 17:39:27.955053 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 27 17:39:27 crc kubenswrapper[4809]: I1127 17:39:27.981864 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 27 17:39:27 crc kubenswrapper[4809]: I1127 17:39:27.992118 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 27 17:39:28 crc kubenswrapper[4809]: I1127 17:39:28.009487 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 27 17:39:28 crc kubenswrapper[4809]: I1127 17:39:28.009533 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 27 17:39:29 crc kubenswrapper[4809]: I1127 17:39:29.982763 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 27 17:39:30 crc kubenswrapper[4809]: I1127 17:39:30.044376 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 17:39:30 crc kubenswrapper[4809]: I1127 17:39:30.049220 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 27 17:39:31 crc kubenswrapper[4809]: I1127 17:39:31.171057 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:31 crc kubenswrapper[4809]: I1127 17:39:31.259166 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-n7gbr"] Nov 27 17:39:31 crc kubenswrapper[4809]: I1127 17:39:31.259427 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" podUID="51a633d0-f7b6-4b1c-b609-7f9684069609" containerName="dnsmasq-dns" containerID="cri-o://aec02d90c49dd7261e47b119faa8e613081399e8e11664be13b98bf493f6f61b" gracePeriod=10 Nov 27 17:39:31 crc kubenswrapper[4809]: I1127 17:39:31.794165 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5bfc74dc5d-snbcn"] Nov 27 17:39:31 crc kubenswrapper[4809]: I1127 17:39:31.822610 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" podUID="51a633d0-f7b6-4b1c-b609-7f9684069609" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: connect: connection refused" Nov 27 17:39:31 crc kubenswrapper[4809]: I1127 17:39:31.908436 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:31 crc kubenswrapper[4809]: I1127 17:39:31.921027 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:32 crc kubenswrapper[4809]: W1127 17:39:32.046292 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a7bab33_46e8_4e18_a0c2_1ca25e22493f.slice/crio-cdc46e5cd244603e39cc9b00789ff91d1178a0d13181d3a7f078319c2e4e055e WatchSource:0}: Error finding container cdc46e5cd244603e39cc9b00789ff91d1178a0d13181d3a7f078319c2e4e055e: Status 404 returned error can't find the container with id cdc46e5cd244603e39cc9b00789ff91d1178a0d13181d3a7f078319c2e4e055e Nov 27 17:39:32 crc kubenswrapper[4809]: I1127 17:39:32.047002 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65c6b6d897-rhmzw"] Nov 27 17:39:32 crc kubenswrapper[4809]: I1127 17:39:32.065301 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bfc74dc5d-snbcn" event={"ID":"fea753ab-13fb-49aa-a430-6c5db50f1e6b","Type":"ContainerStarted","Data":"3fd475121004f727641e0ba4225765ffccda246e29305ca8c0a55d04c8934e45"} Nov 27 17:39:32 crc kubenswrapper[4809]: I1127 17:39:32.068946 4809 generic.go:334] "Generic (PLEG): container finished" podID="51a633d0-f7b6-4b1c-b609-7f9684069609" containerID="aec02d90c49dd7261e47b119faa8e613081399e8e11664be13b98bf493f6f61b" exitCode=0 Nov 27 17:39:32 crc kubenswrapper[4809]: I1127 17:39:32.069014 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" event={"ID":"51a633d0-f7b6-4b1c-b609-7f9684069609","Type":"ContainerDied","Data":"aec02d90c49dd7261e47b119faa8e613081399e8e11664be13b98bf493f6f61b"} Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.078451 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c6b6d897-rhmzw" event={"ID":"1a7bab33-46e8-4e18-a0c2-1ca25e22493f","Type":"ContainerStarted","Data":"cdc46e5cd244603e39cc9b00789ff91d1178a0d13181d3a7f078319c2e4e055e"} Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.220262 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.354152 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-config\") pod \"51a633d0-f7b6-4b1c-b609-7f9684069609\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.354270 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rf9s\" (UniqueName: \"kubernetes.io/projected/51a633d0-f7b6-4b1c-b609-7f9684069609-kube-api-access-5rf9s\") pod \"51a633d0-f7b6-4b1c-b609-7f9684069609\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.354331 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-sb\") pod \"51a633d0-f7b6-4b1c-b609-7f9684069609\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.354371 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-nb\") pod \"51a633d0-f7b6-4b1c-b609-7f9684069609\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.355448 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-swift-storage-0\") pod \"51a633d0-f7b6-4b1c-b609-7f9684069609\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.355622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-svc\") pod \"51a633d0-f7b6-4b1c-b609-7f9684069609\" (UID: \"51a633d0-f7b6-4b1c-b609-7f9684069609\") " Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.361703 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a633d0-f7b6-4b1c-b609-7f9684069609-kube-api-access-5rf9s" (OuterVolumeSpecName: "kube-api-access-5rf9s") pod "51a633d0-f7b6-4b1c-b609-7f9684069609" (UID: "51a633d0-f7b6-4b1c-b609-7f9684069609"). InnerVolumeSpecName "kube-api-access-5rf9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.412102 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-config" (OuterVolumeSpecName: "config") pod "51a633d0-f7b6-4b1c-b609-7f9684069609" (UID: "51a633d0-f7b6-4b1c-b609-7f9684069609"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.419623 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "51a633d0-f7b6-4b1c-b609-7f9684069609" (UID: "51a633d0-f7b6-4b1c-b609-7f9684069609"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.420959 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "51a633d0-f7b6-4b1c-b609-7f9684069609" (UID: "51a633d0-f7b6-4b1c-b609-7f9684069609"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.422724 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51a633d0-f7b6-4b1c-b609-7f9684069609" (UID: "51a633d0-f7b6-4b1c-b609-7f9684069609"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.433160 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "51a633d0-f7b6-4b1c-b609-7f9684069609" (UID: "51a633d0-f7b6-4b1c-b609-7f9684069609"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.458516 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.458560 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.458574 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rf9s\" (UniqueName: \"kubernetes.io/projected/51a633d0-f7b6-4b1c-b609-7f9684069609-kube-api-access-5rf9s\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.458589 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.458605 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:33 crc kubenswrapper[4809]: I1127 17:39:33.458616 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51a633d0-f7b6-4b1c-b609-7f9684069609-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:34 crc kubenswrapper[4809]: E1127 17:39:34.050451 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="596ec651-2088-4459-b0dd-224c67150b5d" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.099295 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c6b6d897-rhmzw" event={"ID":"1a7bab33-46e8-4e18-a0c2-1ca25e22493f","Type":"ContainerStarted","Data":"e042cb842ba8ac688d57379fd73819b1742187d7d5cccd8eb0f9e355ef0f3f40"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.099354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c6b6d897-rhmzw" event={"ID":"1a7bab33-46e8-4e18-a0c2-1ca25e22493f","Type":"ContainerStarted","Data":"e357f88522e86cb9ea9c5446c7984cb15297d76286dd65d3f9f3dc481966457f"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.099426 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.101939 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" event={"ID":"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9","Type":"ContainerStarted","Data":"4fe472344f493135ac374fc9a6490fda63bd00316fd2d82514eaa3f8bb96066a"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.102007 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" event={"ID":"9bc3b7a8-110d-481e-95d5-3d9034c0b2f9","Type":"ContainerStarted","Data":"f46c0b1346be3bf741975d921d3a54813ac5e9aafc654e30f489f376be7ba0d5"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.104305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bfc74dc5d-snbcn" event={"ID":"fea753ab-13fb-49aa-a430-6c5db50f1e6b","Type":"ContainerStarted","Data":"8d78cf4775a1d55039ca6981a0a747ef129a709ecb00fd867bccb3179f89c40c"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.104347 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bfc74dc5d-snbcn" event={"ID":"fea753ab-13fb-49aa-a430-6c5db50f1e6b","Type":"ContainerStarted","Data":"a24b9a0054e47ac0178742342089e5b0da6d500ff25324b9308bf83502bd3c6a"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.104526 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.107331 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-65b4758f4c-5krfp" event={"ID":"af7083a1-8a26-49b0-8b7a-7b412592ed60","Type":"ContainerStarted","Data":"8f9bce2cf294ec5d347d3f2a9ce8c2e98a5f3d5e7ad5ddf7ccbd9e90faa995b1"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.107369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-65b4758f4c-5krfp" event={"ID":"af7083a1-8a26-49b0-8b7a-7b412592ed60","Type":"ContainerStarted","Data":"288026b505c69157fb519f1f040cc03ed3e1112f0d6e4948858a187eef624f7e"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.110557 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596ec651-2088-4459-b0dd-224c67150b5d","Type":"ContainerStarted","Data":"d1ea97d252dae516cc60372d4b582e962b4cb543468583ae0dbf65f9ebf7504c"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.110666 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="ceilometer-central-agent" containerID="cri-o://8f46e7899255c015a731ca5d6df54e769409b95cc0ab6ec66b6c162dea49ca74" gracePeriod=30 Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.110711 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.110760 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="proxy-httpd" containerID="cri-o://d1ea97d252dae516cc60372d4b582e962b4cb543468583ae0dbf65f9ebf7504c" gracePeriod=30 Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.110769 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="ceilometer-notification-agent" containerID="cri-o://ca99702884a990489a3335cd2835a1eec23536939f25917052640d89c57c60ae" gracePeriod=30 Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.126063 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-65c6b6d897-rhmzw" podStartSLOduration=11.126040949 podStartE2EDuration="11.126040949s" podCreationTimestamp="2025-11-27 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:34.124038644 +0000 UTC m=+1809.396496006" watchObservedRunningTime="2025-11-27 17:39:34.126040949 +0000 UTC m=+1809.398498301" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.159287 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" event={"ID":"51a633d0-f7b6-4b1c-b609-7f9684069609","Type":"ContainerDied","Data":"f0e4f95c0d20993839a5eb37c88908770ed6c65a5fe317d199d4b590a345d907"} Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.159357 4809 scope.go:117] "RemoveContainer" containerID="aec02d90c49dd7261e47b119faa8e613081399e8e11664be13b98bf493f6f61b" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.159802 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-n7gbr" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.163663 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-65b4758f4c-5krfp" podStartSLOduration=4.958596561 podStartE2EDuration="15.163640172s" podCreationTimestamp="2025-11-27 17:39:19 +0000 UTC" firstStartedPulling="2025-11-27 17:39:20.994161335 +0000 UTC m=+1796.266618687" lastFinishedPulling="2025-11-27 17:39:31.199204946 +0000 UTC m=+1806.471662298" observedRunningTime="2025-11-27 17:39:34.155312725 +0000 UTC m=+1809.427770077" watchObservedRunningTime="2025-11-27 17:39:34.163640172 +0000 UTC m=+1809.436097534" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.191378 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-84d69cc6b6-vqlb7" podStartSLOduration=5.104982265 podStartE2EDuration="15.191352797s" podCreationTimestamp="2025-11-27 17:39:19 +0000 UTC" firstStartedPulling="2025-11-27 17:39:21.167503053 +0000 UTC m=+1796.439960405" lastFinishedPulling="2025-11-27 17:39:31.253873595 +0000 UTC m=+1806.526330937" observedRunningTime="2025-11-27 17:39:34.176218665 +0000 UTC m=+1809.448676017" watchObservedRunningTime="2025-11-27 17:39:34.191352797 +0000 UTC m=+1809.463810149" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.199163 4809 scope.go:117] "RemoveContainer" containerID="88a26ef4535bdf518029354840c0f3655cdf3e182a723f2311029144ba64dfba" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.259260 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5bfc74dc5d-snbcn" podStartSLOduration=8.259231354 podStartE2EDuration="8.259231354s" podCreationTimestamp="2025-11-27 17:39:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:34.243189898 +0000 UTC m=+1809.515647260" watchObservedRunningTime="2025-11-27 17:39:34.259231354 +0000 UTC m=+1809.531688726" Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.291155 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-n7gbr"] Nov 27 17:39:34 crc kubenswrapper[4809]: I1127 17:39:34.303945 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-n7gbr"] Nov 27 17:39:35 crc kubenswrapper[4809]: I1127 17:39:35.171624 4809 generic.go:334] "Generic (PLEG): container finished" podID="596ec651-2088-4459-b0dd-224c67150b5d" containerID="8f46e7899255c015a731ca5d6df54e769409b95cc0ab6ec66b6c162dea49ca74" exitCode=0 Nov 27 17:39:35 crc kubenswrapper[4809]: I1127 17:39:35.171703 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596ec651-2088-4459-b0dd-224c67150b5d","Type":"ContainerDied","Data":"8f46e7899255c015a731ca5d6df54e769409b95cc0ab6ec66b6c162dea49ca74"} Nov 27 17:39:35 crc kubenswrapper[4809]: I1127 17:39:35.176287 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-t2gnx" event={"ID":"9a3b3964-4402-49a4-86ba-e6e7909082ec","Type":"ContainerStarted","Data":"6bb89e6b1a9e2e608874ec84a5bf22972f53d03e591dc092dd75ef6f78a63cf6"} Nov 27 17:39:35 crc kubenswrapper[4809]: I1127 17:39:35.177024 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:35 crc kubenswrapper[4809]: I1127 17:39:35.237196 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-t2gnx" podStartSLOduration=3.016111375 podStartE2EDuration="49.237154552s" podCreationTimestamp="2025-11-27 17:38:46 +0000 UTC" firstStartedPulling="2025-11-27 17:38:47.705372139 +0000 UTC m=+1762.977829491" lastFinishedPulling="2025-11-27 17:39:33.926415316 +0000 UTC m=+1809.198872668" observedRunningTime="2025-11-27 17:39:35.225151214 +0000 UTC m=+1810.497608566" watchObservedRunningTime="2025-11-27 17:39:35.237154552 +0000 UTC m=+1810.509611904" Nov 27 17:39:35 crc kubenswrapper[4809]: I1127 17:39:35.468586 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a633d0-f7b6-4b1c-b609-7f9684069609" path="/var/lib/kubelet/pods/51a633d0-f7b6-4b1c-b609-7f9684069609/volumes" Nov 27 17:39:36 crc kubenswrapper[4809]: I1127 17:39:36.186890 4809 generic.go:334] "Generic (PLEG): container finished" podID="596ec651-2088-4459-b0dd-224c67150b5d" containerID="ca99702884a990489a3335cd2835a1eec23536939f25917052640d89c57c60ae" exitCode=0 Nov 27 17:39:36 crc kubenswrapper[4809]: I1127 17:39:36.187079 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596ec651-2088-4459-b0dd-224c67150b5d","Type":"ContainerDied","Data":"ca99702884a990489a3335cd2835a1eec23536939f25917052640d89c57c60ae"} Nov 27 17:39:37 crc kubenswrapper[4809]: I1127 17:39:37.457547 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:39:37 crc kubenswrapper[4809]: E1127 17:39:37.458057 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:39:39 crc kubenswrapper[4809]: I1127 17:39:39.214488 4809 generic.go:334] "Generic (PLEG): container finished" podID="9a3b3964-4402-49a4-86ba-e6e7909082ec" containerID="6bb89e6b1a9e2e608874ec84a5bf22972f53d03e591dc092dd75ef6f78a63cf6" exitCode=0 Nov 27 17:39:39 crc kubenswrapper[4809]: I1127 17:39:39.214528 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-t2gnx" event={"ID":"9a3b3964-4402-49a4-86ba-e6e7909082ec","Type":"ContainerDied","Data":"6bb89e6b1a9e2e608874ec84a5bf22972f53d03e591dc092dd75ef6f78a63cf6"} Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.561037 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.711295 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-db-sync-config-data\") pod \"9a3b3964-4402-49a4-86ba-e6e7909082ec\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.711805 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-scripts\") pod \"9a3b3964-4402-49a4-86ba-e6e7909082ec\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.712013 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-config-data\") pod \"9a3b3964-4402-49a4-86ba-e6e7909082ec\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.712785 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-combined-ca-bundle\") pod \"9a3b3964-4402-49a4-86ba-e6e7909082ec\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.712862 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3b3964-4402-49a4-86ba-e6e7909082ec-etc-machine-id\") pod \"9a3b3964-4402-49a4-86ba-e6e7909082ec\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.712906 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrv44\" (UniqueName: \"kubernetes.io/projected/9a3b3964-4402-49a4-86ba-e6e7909082ec-kube-api-access-hrv44\") pod \"9a3b3964-4402-49a4-86ba-e6e7909082ec\" (UID: \"9a3b3964-4402-49a4-86ba-e6e7909082ec\") " Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.713077 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a3b3964-4402-49a4-86ba-e6e7909082ec-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9a3b3964-4402-49a4-86ba-e6e7909082ec" (UID: "9a3b3964-4402-49a4-86ba-e6e7909082ec"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.713952 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3b3964-4402-49a4-86ba-e6e7909082ec-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.717503 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-scripts" (OuterVolumeSpecName: "scripts") pod "9a3b3964-4402-49a4-86ba-e6e7909082ec" (UID: "9a3b3964-4402-49a4-86ba-e6e7909082ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.717973 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a3b3964-4402-49a4-86ba-e6e7909082ec-kube-api-access-hrv44" (OuterVolumeSpecName: "kube-api-access-hrv44") pod "9a3b3964-4402-49a4-86ba-e6e7909082ec" (UID: "9a3b3964-4402-49a4-86ba-e6e7909082ec"). InnerVolumeSpecName "kube-api-access-hrv44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.718334 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9a3b3964-4402-49a4-86ba-e6e7909082ec" (UID: "9a3b3964-4402-49a4-86ba-e6e7909082ec"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.738787 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a3b3964-4402-49a4-86ba-e6e7909082ec" (UID: "9a3b3964-4402-49a4-86ba-e6e7909082ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.760615 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-config-data" (OuterVolumeSpecName: "config-data") pod "9a3b3964-4402-49a4-86ba-e6e7909082ec" (UID: "9a3b3964-4402-49a4-86ba-e6e7909082ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.815886 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.815925 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.815938 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.815952 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrv44\" (UniqueName: \"kubernetes.io/projected/9a3b3964-4402-49a4-86ba-e6e7909082ec-kube-api-access-hrv44\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:40 crc kubenswrapper[4809]: I1127 17:39:40.815964 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3b3964-4402-49a4-86ba-e6e7909082ec-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.232141 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-t2gnx" event={"ID":"9a3b3964-4402-49a4-86ba-e6e7909082ec","Type":"ContainerDied","Data":"e76ed9602e46c1d8479bb657c98a72e04515d23f7b25b2ef5ec8292ec733365d"} Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.232186 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e76ed9602e46c1d8479bb657c98a72e04515d23f7b25b2ef5ec8292ec733365d" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.232262 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-t2gnx" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.508061 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 17:39:41 crc kubenswrapper[4809]: E1127 17:39:41.508499 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a633d0-f7b6-4b1c-b609-7f9684069609" containerName="dnsmasq-dns" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.508515 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a633d0-f7b6-4b1c-b609-7f9684069609" containerName="dnsmasq-dns" Nov 27 17:39:41 crc kubenswrapper[4809]: E1127 17:39:41.508535 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3b3964-4402-49a4-86ba-e6e7909082ec" containerName="cinder-db-sync" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.508543 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3b3964-4402-49a4-86ba-e6e7909082ec" containerName="cinder-db-sync" Nov 27 17:39:41 crc kubenswrapper[4809]: E1127 17:39:41.508562 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a633d0-f7b6-4b1c-b609-7f9684069609" containerName="init" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.508568 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a633d0-f7b6-4b1c-b609-7f9684069609" containerName="init" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.508750 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3b3964-4402-49a4-86ba-e6e7909082ec" containerName="cinder-db-sync" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.508776 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a633d0-f7b6-4b1c-b609-7f9684069609" containerName="dnsmasq-dns" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.510091 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.515802 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2vvhs" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.516036 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.516834 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.517080 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.519470 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.592340 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5x7sr"] Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.594202 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.620634 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5x7sr"] Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.635572 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.635631 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.635760 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn4bl\" (UniqueName: \"kubernetes.io/projected/983737cc-d82d-4d76-a060-b27cb8bcd94d-kube-api-access-rn4bl\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.635785 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.635814 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/983737cc-d82d-4d76-a060-b27cb8bcd94d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.635829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-scripts\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.717060 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.719119 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.723622 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.736867 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-scripts\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.736968 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-config\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737009 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737035 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737090 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jpx9\" (UniqueName: \"kubernetes.io/projected/5ff6795d-3222-4d84-821b-86a303e0e74e-kube-api-access-2jpx9\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737186 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-svc\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737214 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn4bl\" (UniqueName: \"kubernetes.io/projected/983737cc-d82d-4d76-a060-b27cb8bcd94d-kube-api-access-rn4bl\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737236 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/983737cc-d82d-4d76-a060-b27cb8bcd94d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.737251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.743267 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.744238 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/983737cc-d82d-4d76-a060-b27cb8bcd94d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.747716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-scripts\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.748124 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.767667 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.768495 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.771086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn4bl\" (UniqueName: \"kubernetes.io/projected/983737cc-d82d-4d76-a060-b27cb8bcd94d-kube-api-access-rn4bl\") pod \"cinder-scheduler-0\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.827949 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839488 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfq5s\" (UniqueName: \"kubernetes.io/projected/4414a560-f463-404a-8513-0113d19423e4-kube-api-access-mfq5s\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839635 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data-custom\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4414a560-f463-404a-8513-0113d19423e4-logs\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839707 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4414a560-f463-404a-8513-0113d19423e4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839775 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-config\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839807 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839863 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jpx9\" (UniqueName: \"kubernetes.io/projected/5ff6795d-3222-4d84-821b-86a303e0e74e-kube-api-access-2jpx9\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839936 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.839956 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.840006 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.840026 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-scripts\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.840048 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-svc\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.841018 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-svc\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.845057 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.845281 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-config\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.845951 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.845990 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.867833 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jpx9\" (UniqueName: \"kubernetes.io/projected/5ff6795d-3222-4d84-821b-86a303e0e74e-kube-api-access-2jpx9\") pod \"dnsmasq-dns-5784cf869f-5x7sr\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.914021 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.941334 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfq5s\" (UniqueName: \"kubernetes.io/projected/4414a560-f463-404a-8513-0113d19423e4-kube-api-access-mfq5s\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.941386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data-custom\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.941410 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4414a560-f463-404a-8513-0113d19423e4-logs\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.941435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4414a560-f463-404a-8513-0113d19423e4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.941476 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.941535 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.941575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-scripts\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.942651 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4414a560-f463-404a-8513-0113d19423e4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.942887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4414a560-f463-404a-8513-0113d19423e4-logs\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.955160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-scripts\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.955501 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.957422 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data-custom\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.961820 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:41 crc kubenswrapper[4809]: I1127 17:39:41.970396 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfq5s\" (UniqueName: \"kubernetes.io/projected/4414a560-f463-404a-8513-0113d19423e4-kube-api-access-mfq5s\") pod \"cinder-api-0\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " pod="openstack/cinder-api-0" Nov 27 17:39:42 crc kubenswrapper[4809]: I1127 17:39:42.045828 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 17:39:42 crc kubenswrapper[4809]: I1127 17:39:42.433657 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5x7sr"] Nov 27 17:39:42 crc kubenswrapper[4809]: I1127 17:39:42.521530 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 17:39:42 crc kubenswrapper[4809]: W1127 17:39:42.529329 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod983737cc_d82d_4d76_a060_b27cb8bcd94d.slice/crio-299cdd87e424cf871fde98ea449c49ce26876a78710d023a8f7f3948527ea278 WatchSource:0}: Error finding container 299cdd87e424cf871fde98ea449c49ce26876a78710d023a8f7f3948527ea278: Status 404 returned error can't find the container with id 299cdd87e424cf871fde98ea449c49ce26876a78710d023a8f7f3948527ea278 Nov 27 17:39:42 crc kubenswrapper[4809]: I1127 17:39:42.702212 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 17:39:42 crc kubenswrapper[4809]: W1127 17:39:42.763193 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4414a560_f463_404a_8513_0113d19423e4.slice/crio-cf3e73a413358d4c63534908d35cdde69f56e19bcd6cd72ebe980d63d6d4cf19 WatchSource:0}: Error finding container cf3e73a413358d4c63534908d35cdde69f56e19bcd6cd72ebe980d63d6d4cf19: Status 404 returned error can't find the container with id cf3e73a413358d4c63534908d35cdde69f56e19bcd6cd72ebe980d63d6d4cf19 Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.272084 4809 generic.go:334] "Generic (PLEG): container finished" podID="5ff6795d-3222-4d84-821b-86a303e0e74e" containerID="41a970cdf1341fe0baa0a36ca62e94be78e1d62997846ac4a45ae696e3c87978" exitCode=0 Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.272201 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" event={"ID":"5ff6795d-3222-4d84-821b-86a303e0e74e","Type":"ContainerDied","Data":"41a970cdf1341fe0baa0a36ca62e94be78e1d62997846ac4a45ae696e3c87978"} Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.272444 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" event={"ID":"5ff6795d-3222-4d84-821b-86a303e0e74e","Type":"ContainerStarted","Data":"93cb8a84c482744cbc1dcbbae348d8eb2101f4ce7b86999c0962dbe1b153ffbe"} Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.275233 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"983737cc-d82d-4d76-a060-b27cb8bcd94d","Type":"ContainerStarted","Data":"299cdd87e424cf871fde98ea449c49ce26876a78710d023a8f7f3948527ea278"} Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.277453 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4414a560-f463-404a-8513-0113d19423e4","Type":"ContainerStarted","Data":"cf3e73a413358d4c63534908d35cdde69f56e19bcd6cd72ebe980d63d6d4cf19"} Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.281979 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.441113 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5bfc74dc5d-snbcn" Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.506259 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6688bc68c4-s95bj"] Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.509234 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6688bc68c4-s95bj" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api" containerID="cri-o://4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9" gracePeriod=30 Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.508804 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6688bc68c4-s95bj" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api-log" containerID="cri-o://13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718" gracePeriod=30 Nov 27 17:39:43 crc kubenswrapper[4809]: I1127 17:39:43.664426 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 27 17:39:44 crc kubenswrapper[4809]: I1127 17:39:44.292061 4809 generic.go:334] "Generic (PLEG): container finished" podID="c406df20-1687-4079-8b4b-8466a334e9a8" containerID="13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718" exitCode=143 Nov 27 17:39:44 crc kubenswrapper[4809]: I1127 17:39:44.292238 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6688bc68c4-s95bj" event={"ID":"c406df20-1687-4079-8b4b-8466a334e9a8","Type":"ContainerDied","Data":"13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718"} Nov 27 17:39:44 crc kubenswrapper[4809]: I1127 17:39:44.293729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4414a560-f463-404a-8513-0113d19423e4","Type":"ContainerStarted","Data":"2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e"} Nov 27 17:39:44 crc kubenswrapper[4809]: I1127 17:39:44.297396 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" event={"ID":"5ff6795d-3222-4d84-821b-86a303e0e74e","Type":"ContainerStarted","Data":"122d6c5478b056259670d4dc4929caf6a5533924f924d753ef2b465d4788cd16"} Nov 27 17:39:44 crc kubenswrapper[4809]: I1127 17:39:44.297444 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:44 crc kubenswrapper[4809]: I1127 17:39:44.321309 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" podStartSLOduration=3.321287625 podStartE2EDuration="3.321287625s" podCreationTimestamp="2025-11-27 17:39:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:44.313455621 +0000 UTC m=+1819.585912973" watchObservedRunningTime="2025-11-27 17:39:44.321287625 +0000 UTC m=+1819.593744977" Nov 27 17:39:45 crc kubenswrapper[4809]: I1127 17:39:45.305775 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4414a560-f463-404a-8513-0113d19423e4","Type":"ContainerStarted","Data":"8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b"} Nov 27 17:39:45 crc kubenswrapper[4809]: I1127 17:39:45.306333 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 27 17:39:45 crc kubenswrapper[4809]: I1127 17:39:45.305931 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4414a560-f463-404a-8513-0113d19423e4" containerName="cinder-api" containerID="cri-o://8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b" gracePeriod=30 Nov 27 17:39:45 crc kubenswrapper[4809]: I1127 17:39:45.305876 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4414a560-f463-404a-8513-0113d19423e4" containerName="cinder-api-log" containerID="cri-o://2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e" gracePeriod=30 Nov 27 17:39:45 crc kubenswrapper[4809]: I1127 17:39:45.311689 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"983737cc-d82d-4d76-a060-b27cb8bcd94d","Type":"ContainerStarted","Data":"11e20e37adde989f6d2996800e67f6f3713a090171d816faf51ea99512f137bf"} Nov 27 17:39:45 crc kubenswrapper[4809]: I1127 17:39:45.311750 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"983737cc-d82d-4d76-a060-b27cb8bcd94d","Type":"ContainerStarted","Data":"4361680b2d04b78af6a69f3c07134a57666e9e73d712670abf2504e350e9a498"} Nov 27 17:39:45 crc kubenswrapper[4809]: I1127 17:39:45.353028 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.353003137 podStartE2EDuration="4.353003137s" podCreationTimestamp="2025-11-27 17:39:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:45.338039439 +0000 UTC m=+1820.610496801" watchObservedRunningTime="2025-11-27 17:39:45.353003137 +0000 UTC m=+1820.625460489" Nov 27 17:39:45 crc kubenswrapper[4809]: I1127 17:39:45.374225 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.245098761 podStartE2EDuration="4.374207224s" podCreationTimestamp="2025-11-27 17:39:41 +0000 UTC" firstStartedPulling="2025-11-27 17:39:42.533646578 +0000 UTC m=+1817.806103930" lastFinishedPulling="2025-11-27 17:39:43.662755041 +0000 UTC m=+1818.935212393" observedRunningTime="2025-11-27 17:39:45.356295457 +0000 UTC m=+1820.628752809" watchObservedRunningTime="2025-11-27 17:39:45.374207224 +0000 UTC m=+1820.646664566" Nov 27 17:39:45 crc kubenswrapper[4809]: I1127 17:39:45.912533 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.047802 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4414a560-f463-404a-8513-0113d19423e4-logs\") pod \"4414a560-f463-404a-8513-0113d19423e4\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.047894 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-combined-ca-bundle\") pod \"4414a560-f463-404a-8513-0113d19423e4\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.048087 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-scripts\") pod \"4414a560-f463-404a-8513-0113d19423e4\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.048128 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfq5s\" (UniqueName: \"kubernetes.io/projected/4414a560-f463-404a-8513-0113d19423e4-kube-api-access-mfq5s\") pod \"4414a560-f463-404a-8513-0113d19423e4\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.048145 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data\") pod \"4414a560-f463-404a-8513-0113d19423e4\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.048203 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4414a560-f463-404a-8513-0113d19423e4-etc-machine-id\") pod \"4414a560-f463-404a-8513-0113d19423e4\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.048222 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data-custom\") pod \"4414a560-f463-404a-8513-0113d19423e4\" (UID: \"4414a560-f463-404a-8513-0113d19423e4\") " Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.049821 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4414a560-f463-404a-8513-0113d19423e4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4414a560-f463-404a-8513-0113d19423e4" (UID: "4414a560-f463-404a-8513-0113d19423e4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.049948 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4414a560-f463-404a-8513-0113d19423e4-logs" (OuterVolumeSpecName: "logs") pod "4414a560-f463-404a-8513-0113d19423e4" (UID: "4414a560-f463-404a-8513-0113d19423e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.055399 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-scripts" (OuterVolumeSpecName: "scripts") pod "4414a560-f463-404a-8513-0113d19423e4" (UID: "4414a560-f463-404a-8513-0113d19423e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.055899 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4414a560-f463-404a-8513-0113d19423e4" (UID: "4414a560-f463-404a-8513-0113d19423e4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.057488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4414a560-f463-404a-8513-0113d19423e4-kube-api-access-mfq5s" (OuterVolumeSpecName: "kube-api-access-mfq5s") pod "4414a560-f463-404a-8513-0113d19423e4" (UID: "4414a560-f463-404a-8513-0113d19423e4"). InnerVolumeSpecName "kube-api-access-mfq5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.088943 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4414a560-f463-404a-8513-0113d19423e4" (UID: "4414a560-f463-404a-8513-0113d19423e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.111632 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data" (OuterVolumeSpecName: "config-data") pod "4414a560-f463-404a-8513-0113d19423e4" (UID: "4414a560-f463-404a-8513-0113d19423e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.149966 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.150002 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfq5s\" (UniqueName: \"kubernetes.io/projected/4414a560-f463-404a-8513-0113d19423e4-kube-api-access-mfq5s\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.150017 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.150029 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4414a560-f463-404a-8513-0113d19423e4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.150039 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.150050 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4414a560-f463-404a-8513-0113d19423e4-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.150058 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4414a560-f463-404a-8513-0113d19423e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.329999 4809 generic.go:334] "Generic (PLEG): container finished" podID="4414a560-f463-404a-8513-0113d19423e4" containerID="8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b" exitCode=0 Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.330036 4809 generic.go:334] "Generic (PLEG): container finished" podID="4414a560-f463-404a-8513-0113d19423e4" containerID="2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e" exitCode=143 Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.330058 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4414a560-f463-404a-8513-0113d19423e4","Type":"ContainerDied","Data":"8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b"} Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.330115 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.330161 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4414a560-f463-404a-8513-0113d19423e4","Type":"ContainerDied","Data":"2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e"} Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.330180 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4414a560-f463-404a-8513-0113d19423e4","Type":"ContainerDied","Data":"cf3e73a413358d4c63534908d35cdde69f56e19bcd6cd72ebe980d63d6d4cf19"} Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.330213 4809 scope.go:117] "RemoveContainer" containerID="8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.357147 4809 scope.go:117] "RemoveContainer" containerID="2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.379433 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.398664 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.398795 4809 scope.go:117] "RemoveContainer" containerID="8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b" Nov 27 17:39:46 crc kubenswrapper[4809]: E1127 17:39:46.399587 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b\": container with ID starting with 8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b not found: ID does not exist" containerID="8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.399624 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b"} err="failed to get container status \"8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b\": rpc error: code = NotFound desc = could not find container \"8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b\": container with ID starting with 8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b not found: ID does not exist" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.399653 4809 scope.go:117] "RemoveContainer" containerID="2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e" Nov 27 17:39:46 crc kubenswrapper[4809]: E1127 17:39:46.400301 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e\": container with ID starting with 2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e not found: ID does not exist" containerID="2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.400334 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e"} err="failed to get container status \"2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e\": rpc error: code = NotFound desc = could not find container \"2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e\": container with ID starting with 2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e not found: ID does not exist" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.400372 4809 scope.go:117] "RemoveContainer" containerID="8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.400951 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b"} err="failed to get container status \"8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b\": rpc error: code = NotFound desc = could not find container \"8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b\": container with ID starting with 8f49b09473e447fdfb5540ec7c33da493c497da70ecc7ee22e8497146df2d98b not found: ID does not exist" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.400991 4809 scope.go:117] "RemoveContainer" containerID="2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.401363 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e"} err="failed to get container status \"2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e\": rpc error: code = NotFound desc = could not find container \"2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e\": container with ID starting with 2a938865db23417991a37a4fa3bb33f640e48eeea6d7179b00636990d9bbde4e not found: ID does not exist" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.407321 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 27 17:39:46 crc kubenswrapper[4809]: E1127 17:39:46.408726 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4414a560-f463-404a-8513-0113d19423e4" containerName="cinder-api" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.408770 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4414a560-f463-404a-8513-0113d19423e4" containerName="cinder-api" Nov 27 17:39:46 crc kubenswrapper[4809]: E1127 17:39:46.408791 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4414a560-f463-404a-8513-0113d19423e4" containerName="cinder-api-log" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.408799 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4414a560-f463-404a-8513-0113d19423e4" containerName="cinder-api-log" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.409046 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4414a560-f463-404a-8513-0113d19423e4" containerName="cinder-api" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.409072 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4414a560-f463-404a-8513-0113d19423e4" containerName="cinder-api-log" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.410323 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.412857 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.413057 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.413174 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.418395 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.558309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-scripts\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.558359 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.558395 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.558454 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.558488 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-config-data\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.558517 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-logs\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.558566 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.558618 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.558641 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcw82\" (UniqueName: \"kubernetes.io/projected/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-kube-api-access-hcw82\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.656285 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6688bc68c4-s95bj" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.172:9311/healthcheck\": read tcp 10.217.0.2:35130->10.217.0.172:9311: read: connection reset by peer" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.656356 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6688bc68c4-s95bj" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.172:9311/healthcheck\": read tcp 10.217.0.2:35124->10.217.0.172:9311: read: connection reset by peer" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.660630 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcw82\" (UniqueName: \"kubernetes.io/projected/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-kube-api-access-hcw82\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.660881 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-scripts\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.661005 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.661093 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.661160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.661275 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.661380 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-config-data\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.661459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-logs\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.661578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.661696 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.661929 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-logs\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.667039 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.667489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-scripts\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.667768 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.668531 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-config-data\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.669042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.671617 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.683065 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcw82\" (UniqueName: \"kubernetes.io/projected/1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6-kube-api-access-hcw82\") pod \"cinder-api-0\" (UID: \"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6\") " pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.797919 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.829374 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 27 17:39:46 crc kubenswrapper[4809]: I1127 17:39:46.837793 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.086642 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.253787 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 17:39:47 crc kubenswrapper[4809]: W1127 17:39:47.257228 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1adad9c7_0fbf_4ae3_ad2c_f3177a66b1c6.slice/crio-a8cc37e162a42f7ea1df19c1d5477590c629511684cc1d0fc1974184aa1a48cd WatchSource:0}: Error finding container a8cc37e162a42f7ea1df19c1d5477590c629511684cc1d0fc1974184aa1a48cd: Status 404 returned error can't find the container with id a8cc37e162a42f7ea1df19c1d5477590c629511684cc1d0fc1974184aa1a48cd Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.273981 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data\") pod \"c406df20-1687-4079-8b4b-8466a334e9a8\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.274082 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-combined-ca-bundle\") pod \"c406df20-1687-4079-8b4b-8466a334e9a8\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.274222 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data-custom\") pod \"c406df20-1687-4079-8b4b-8466a334e9a8\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.274246 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lvdj\" (UniqueName: \"kubernetes.io/projected/c406df20-1687-4079-8b4b-8466a334e9a8-kube-api-access-2lvdj\") pod \"c406df20-1687-4079-8b4b-8466a334e9a8\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.274429 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c406df20-1687-4079-8b4b-8466a334e9a8-logs\") pod \"c406df20-1687-4079-8b4b-8466a334e9a8\" (UID: \"c406df20-1687-4079-8b4b-8466a334e9a8\") " Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.274800 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c406df20-1687-4079-8b4b-8466a334e9a8-logs" (OuterVolumeSpecName: "logs") pod "c406df20-1687-4079-8b4b-8466a334e9a8" (UID: "c406df20-1687-4079-8b4b-8466a334e9a8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.275157 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c406df20-1687-4079-8b4b-8466a334e9a8-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.283907 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c406df20-1687-4079-8b4b-8466a334e9a8" (UID: "c406df20-1687-4079-8b4b-8466a334e9a8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.283950 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c406df20-1687-4079-8b4b-8466a334e9a8-kube-api-access-2lvdj" (OuterVolumeSpecName: "kube-api-access-2lvdj") pod "c406df20-1687-4079-8b4b-8466a334e9a8" (UID: "c406df20-1687-4079-8b4b-8466a334e9a8"). InnerVolumeSpecName "kube-api-access-2lvdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.304101 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c406df20-1687-4079-8b4b-8466a334e9a8" (UID: "c406df20-1687-4079-8b4b-8466a334e9a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.329256 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data" (OuterVolumeSpecName: "config-data") pod "c406df20-1687-4079-8b4b-8466a334e9a8" (UID: "c406df20-1687-4079-8b4b-8466a334e9a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.339656 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6","Type":"ContainerStarted","Data":"a8cc37e162a42f7ea1df19c1d5477590c629511684cc1d0fc1974184aa1a48cd"} Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.342284 4809 generic.go:334] "Generic (PLEG): container finished" podID="c406df20-1687-4079-8b4b-8466a334e9a8" containerID="4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9" exitCode=0 Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.342365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6688bc68c4-s95bj" event={"ID":"c406df20-1687-4079-8b4b-8466a334e9a8","Type":"ContainerDied","Data":"4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9"} Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.342400 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6688bc68c4-s95bj" event={"ID":"c406df20-1687-4079-8b4b-8466a334e9a8","Type":"ContainerDied","Data":"456cef95bf8eddad75cc79095d63d87fca5c1478dd0cd8694b58838668737057"} Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.342422 4809 scope.go:117] "RemoveContainer" containerID="4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.342536 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6688bc68c4-s95bj" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.374166 4809 scope.go:117] "RemoveContainer" containerID="13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.376568 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.376591 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.376600 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c406df20-1687-4079-8b4b-8466a334e9a8-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.376609 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lvdj\" (UniqueName: \"kubernetes.io/projected/c406df20-1687-4079-8b4b-8466a334e9a8-kube-api-access-2lvdj\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.381887 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6688bc68c4-s95bj"] Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.390839 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6688bc68c4-s95bj"] Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.392431 4809 scope.go:117] "RemoveContainer" containerID="4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9" Nov 27 17:39:47 crc kubenswrapper[4809]: E1127 17:39:47.393145 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9\": container with ID starting with 4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9 not found: ID does not exist" containerID="4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.393197 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9"} err="failed to get container status \"4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9\": rpc error: code = NotFound desc = could not find container \"4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9\": container with ID starting with 4e89ac2b0aa1bff9da9a9239182e45c29ea5602e50453fec2863092bc75500f9 not found: ID does not exist" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.393232 4809 scope.go:117] "RemoveContainer" containerID="13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718" Nov 27 17:39:47 crc kubenswrapper[4809]: E1127 17:39:47.393502 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718\": container with ID starting with 13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718 not found: ID does not exist" containerID="13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.393536 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718"} err="failed to get container status \"13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718\": rpc error: code = NotFound desc = could not find container \"13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718\": container with ID starting with 13fdd299dd7fd5b19416d60b80f7512335201717514e1d8af663b1f72dfc4718 not found: ID does not exist" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.469517 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4414a560-f463-404a-8513-0113d19423e4" path="/var/lib/kubelet/pods/4414a560-f463-404a-8513-0113d19423e4/volumes" Nov 27 17:39:47 crc kubenswrapper[4809]: I1127 17:39:47.470495 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" path="/var/lib/kubelet/pods/c406df20-1687-4079-8b4b-8466a334e9a8/volumes" Nov 27 17:39:48 crc kubenswrapper[4809]: I1127 17:39:48.354456 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6","Type":"ContainerStarted","Data":"d95e3bbc630ca39e083cb8a20bcc159d0faad71408dcc6492a92f59d395daaa1"} Nov 27 17:39:48 crc kubenswrapper[4809]: I1127 17:39:48.355068 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 27 17:39:48 crc kubenswrapper[4809]: I1127 17:39:48.355081 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6","Type":"ContainerStarted","Data":"b94459404c899e43feac9177183553b7a549bc72307f865203a338f205b88e7b"} Nov 27 17:39:48 crc kubenswrapper[4809]: I1127 17:39:48.371857 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.371843673 podStartE2EDuration="2.371843673s" podCreationTimestamp="2025-11-27 17:39:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:48.371286158 +0000 UTC m=+1823.643743520" watchObservedRunningTime="2025-11-27 17:39:48.371843673 +0000 UTC m=+1823.644301025" Nov 27 17:39:48 crc kubenswrapper[4809]: I1127 17:39:48.458233 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:39:48 crc kubenswrapper[4809]: E1127 17:39:48.458557 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:39:51 crc kubenswrapper[4809]: I1127 17:39:51.230180 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:51 crc kubenswrapper[4809]: I1127 17:39:51.239448 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:51 crc kubenswrapper[4809]: I1127 17:39:51.240623 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-556f88f66d-dk28p" Nov 27 17:39:51 crc kubenswrapper[4809]: I1127 17:39:51.523846 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-856cbdcf9c-lws7x" Nov 27 17:39:51 crc kubenswrapper[4809]: I1127 17:39:51.915898 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:39:51 crc kubenswrapper[4809]: I1127 17:39:51.974566 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-qslw2"] Nov 27 17:39:51 crc kubenswrapper[4809]: I1127 17:39:51.975787 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" podUID="b1890991-c957-47ca-89d4-307e357808d4" containerName="dnsmasq-dns" containerID="cri-o://4740da13b9f9a3dab0f9167c31fb3d7f2507674dae05217ffc335c739389f02e" gracePeriod=10 Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.131599 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.194080 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.398664 4809 generic.go:334] "Generic (PLEG): container finished" podID="b1890991-c957-47ca-89d4-307e357808d4" containerID="4740da13b9f9a3dab0f9167c31fb3d7f2507674dae05217ffc335c739389f02e" exitCode=0 Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.398814 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" event={"ID":"b1890991-c957-47ca-89d4-307e357808d4","Type":"ContainerDied","Data":"4740da13b9f9a3dab0f9167c31fb3d7f2507674dae05217ffc335c739389f02e"} Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.398972 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerName="cinder-scheduler" containerID="cri-o://4361680b2d04b78af6a69f3c07134a57666e9e73d712670abf2504e350e9a498" gracePeriod=30 Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.399001 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerName="probe" containerID="cri-o://11e20e37adde989f6d2996800e67f6f3713a090171d816faf51ea99512f137bf" gracePeriod=30 Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.518206 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.676539 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-sb\") pod \"b1890991-c957-47ca-89d4-307e357808d4\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.676618 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-config\") pod \"b1890991-c957-47ca-89d4-307e357808d4\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.676669 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-swift-storage-0\") pod \"b1890991-c957-47ca-89d4-307e357808d4\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.676751 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h72w4\" (UniqueName: \"kubernetes.io/projected/b1890991-c957-47ca-89d4-307e357808d4-kube-api-access-h72w4\") pod \"b1890991-c957-47ca-89d4-307e357808d4\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.676848 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-nb\") pod \"b1890991-c957-47ca-89d4-307e357808d4\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.676868 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-svc\") pod \"b1890991-c957-47ca-89d4-307e357808d4\" (UID: \"b1890991-c957-47ca-89d4-307e357808d4\") " Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.682774 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1890991-c957-47ca-89d4-307e357808d4-kube-api-access-h72w4" (OuterVolumeSpecName: "kube-api-access-h72w4") pod "b1890991-c957-47ca-89d4-307e357808d4" (UID: "b1890991-c957-47ca-89d4-307e357808d4"). InnerVolumeSpecName "kube-api-access-h72w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.725261 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b1890991-c957-47ca-89d4-307e357808d4" (UID: "b1890991-c957-47ca-89d4-307e357808d4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.725916 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b1890991-c957-47ca-89d4-307e357808d4" (UID: "b1890991-c957-47ca-89d4-307e357808d4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.732331 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-config" (OuterVolumeSpecName: "config") pod "b1890991-c957-47ca-89d4-307e357808d4" (UID: "b1890991-c957-47ca-89d4-307e357808d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.737468 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b1890991-c957-47ca-89d4-307e357808d4" (UID: "b1890991-c957-47ca-89d4-307e357808d4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.740716 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b1890991-c957-47ca-89d4-307e357808d4" (UID: "b1890991-c957-47ca-89d4-307e357808d4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.779996 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.780038 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.780049 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.780060 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h72w4\" (UniqueName: \"kubernetes.io/projected/b1890991-c957-47ca-89d4-307e357808d4-kube-api-access-h72w4\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.780072 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:52 crc kubenswrapper[4809]: I1127 17:39:52.780082 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1890991-c957-47ca-89d4-307e357808d4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:53 crc kubenswrapper[4809]: I1127 17:39:53.410262 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" event={"ID":"b1890991-c957-47ca-89d4-307e357808d4","Type":"ContainerDied","Data":"573381ea3f7970cae0be63e0c7bcd4ef310d777b25754e47af3aecd4d1a71ec2"} Nov 27 17:39:53 crc kubenswrapper[4809]: I1127 17:39:53.410352 4809 scope.go:117] "RemoveContainer" containerID="4740da13b9f9a3dab0f9167c31fb3d7f2507674dae05217ffc335c739389f02e" Nov 27 17:39:53 crc kubenswrapper[4809]: I1127 17:39:53.410285 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-qslw2" Nov 27 17:39:53 crc kubenswrapper[4809]: I1127 17:39:53.412631 4809 generic.go:334] "Generic (PLEG): container finished" podID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerID="11e20e37adde989f6d2996800e67f6f3713a090171d816faf51ea99512f137bf" exitCode=0 Nov 27 17:39:53 crc kubenswrapper[4809]: I1127 17:39:53.412695 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"983737cc-d82d-4d76-a060-b27cb8bcd94d","Type":"ContainerDied","Data":"11e20e37adde989f6d2996800e67f6f3713a090171d816faf51ea99512f137bf"} Nov 27 17:39:53 crc kubenswrapper[4809]: I1127 17:39:53.431919 4809 scope.go:117] "RemoveContainer" containerID="a452e9686dda99a1da061cb064f877094c9ade39be9a185d42144a0a98233917" Nov 27 17:39:53 crc kubenswrapper[4809]: I1127 17:39:53.450998 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-qslw2"] Nov 27 17:39:53 crc kubenswrapper[4809]: I1127 17:39:53.470991 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-qslw2"] Nov 27 17:39:53 crc kubenswrapper[4809]: I1127 17:39:53.951601 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-65c6b6d897-rhmzw" Nov 27 17:39:54 crc kubenswrapper[4809]: I1127 17:39:54.011182 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cffcf7bfb-7p9tr"] Nov 27 17:39:54 crc kubenswrapper[4809]: I1127 17:39:54.011489 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cffcf7bfb-7p9tr" podUID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerName="neutron-api" containerID="cri-o://9035331522b1447565598f7473b7ab808d2ec9ca68cd09992c867d8b5192020b" gracePeriod=30 Nov 27 17:39:54 crc kubenswrapper[4809]: I1127 17:39:54.011973 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cffcf7bfb-7p9tr" podUID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerName="neutron-httpd" containerID="cri-o://51d28b59af8a554b584b8f720373c960b56ded2c93ae0456729f8b49d97db81c" gracePeriod=30 Nov 27 17:39:54 crc kubenswrapper[4809]: I1127 17:39:54.423694 4809 generic.go:334] "Generic (PLEG): container finished" podID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerID="51d28b59af8a554b584b8f720373c960b56ded2c93ae0456729f8b49d97db81c" exitCode=0 Nov 27 17:39:54 crc kubenswrapper[4809]: I1127 17:39:54.423770 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffcf7bfb-7p9tr" event={"ID":"013eaaea-f9de-4e5c-9647-a5df367b7709","Type":"ContainerDied","Data":"51d28b59af8a554b584b8f720373c960b56ded2c93ae0456729f8b49d97db81c"} Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.408204 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 27 17:39:55 crc kubenswrapper[4809]: E1127 17:39:55.409002 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1890991-c957-47ca-89d4-307e357808d4" containerName="dnsmasq-dns" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.409025 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1890991-c957-47ca-89d4-307e357808d4" containerName="dnsmasq-dns" Nov 27 17:39:55 crc kubenswrapper[4809]: E1127 17:39:55.409037 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api-log" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.409045 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api-log" Nov 27 17:39:55 crc kubenswrapper[4809]: E1127 17:39:55.409060 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1890991-c957-47ca-89d4-307e357808d4" containerName="init" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.409067 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1890991-c957-47ca-89d4-307e357808d4" containerName="init" Nov 27 17:39:55 crc kubenswrapper[4809]: E1127 17:39:55.409105 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.409117 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.409327 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1890991-c957-47ca-89d4-307e357808d4" containerName="dnsmasq-dns" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.409343 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.409356 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c406df20-1687-4079-8b4b-8466a334e9a8" containerName="barbican-api-log" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.410113 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.413763 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.413773 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-4q6jz" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.414160 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.440976 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.448524 4809 generic.go:334] "Generic (PLEG): container finished" podID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerID="4361680b2d04b78af6a69f3c07134a57666e9e73d712670abf2504e350e9a498" exitCode=0 Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.448574 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"983737cc-d82d-4d76-a060-b27cb8bcd94d","Type":"ContainerDied","Data":"4361680b2d04b78af6a69f3c07134a57666e9e73d712670abf2504e350e9a498"} Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.470377 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1890991-c957-47ca-89d4-307e357808d4" path="/var/lib/kubelet/pods/b1890991-c957-47ca-89d4-307e357808d4/volumes" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.556949 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9zwv\" (UniqueName: \"kubernetes.io/projected/79db49b5-8439-451b-936d-c4021dc7567e-kube-api-access-v9zwv\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.557086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.557135 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.557205 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config-secret\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.664103 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.665677 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.665588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.665887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config-secret\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.666277 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9zwv\" (UniqueName: \"kubernetes.io/projected/79db49b5-8439-451b-936d-c4021dc7567e-kube-api-access-v9zwv\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.704963 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9zwv\" (UniqueName: \"kubernetes.io/projected/79db49b5-8439-451b-936d-c4021dc7567e-kube-api-access-v9zwv\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.714431 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config-secret\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.716197 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.807381 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.808267 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.817012 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.856003 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.857707 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.883757 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 17:39:55 crc kubenswrapper[4809]: E1127 17:39:55.944241 4809 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 27 17:39:55 crc kubenswrapper[4809]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_79db49b5-8439-451b-936d-c4021dc7567e_0(1e98675ef7e0ff2bc6709d0445ed64698a01851fe9b5d36c804a9b16aa2741d6): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1e98675ef7e0ff2bc6709d0445ed64698a01851fe9b5d36c804a9b16aa2741d6" Netns:"/var/run/netns/fa6a73bd-9b71-46f1-9fa1-07bd0e49c79b" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=1e98675ef7e0ff2bc6709d0445ed64698a01851fe9b5d36c804a9b16aa2741d6;K8S_POD_UID=79db49b5-8439-451b-936d-c4021dc7567e" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/79db49b5-8439-451b-936d-c4021dc7567e]: expected pod UID "79db49b5-8439-451b-936d-c4021dc7567e" but got "090560a3-12be-4a67-b507-0c7764053bf2" from Kube API Nov 27 17:39:55 crc kubenswrapper[4809]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 27 17:39:55 crc kubenswrapper[4809]: > Nov 27 17:39:55 crc kubenswrapper[4809]: E1127 17:39:55.944310 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 27 17:39:55 crc kubenswrapper[4809]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_79db49b5-8439-451b-936d-c4021dc7567e_0(1e98675ef7e0ff2bc6709d0445ed64698a01851fe9b5d36c804a9b16aa2741d6): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1e98675ef7e0ff2bc6709d0445ed64698a01851fe9b5d36c804a9b16aa2741d6" Netns:"/var/run/netns/fa6a73bd-9b71-46f1-9fa1-07bd0e49c79b" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=1e98675ef7e0ff2bc6709d0445ed64698a01851fe9b5d36c804a9b16aa2741d6;K8S_POD_UID=79db49b5-8439-451b-936d-c4021dc7567e" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/79db49b5-8439-451b-936d-c4021dc7567e]: expected pod UID "79db49b5-8439-451b-936d-c4021dc7567e" but got "090560a3-12be-4a67-b507-0c7764053bf2" from Kube API Nov 27 17:39:55 crc kubenswrapper[4809]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 27 17:39:55 crc kubenswrapper[4809]: > pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.973917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhjcx\" (UniqueName: \"kubernetes.io/projected/090560a3-12be-4a67-b507-0c7764053bf2-kube-api-access-lhjcx\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.974364 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/090560a3-12be-4a67-b507-0c7764053bf2-openstack-config\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.974423 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/090560a3-12be-4a67-b507-0c7764053bf2-openstack-config-secret\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:55 crc kubenswrapper[4809]: I1127 17:39:55.974702 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/090560a3-12be-4a67-b507-0c7764053bf2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.006268 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.077222 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/090560a3-12be-4a67-b507-0c7764053bf2-openstack-config\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.077272 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/090560a3-12be-4a67-b507-0c7764053bf2-openstack-config-secret\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.077340 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/090560a3-12be-4a67-b507-0c7764053bf2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.077475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhjcx\" (UniqueName: \"kubernetes.io/projected/090560a3-12be-4a67-b507-0c7764053bf2-kube-api-access-lhjcx\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.078560 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/090560a3-12be-4a67-b507-0c7764053bf2-openstack-config\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.084135 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/090560a3-12be-4a67-b507-0c7764053bf2-openstack-config-secret\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.085160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/090560a3-12be-4a67-b507-0c7764053bf2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.095856 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhjcx\" (UniqueName: \"kubernetes.io/projected/090560a3-12be-4a67-b507-0c7764053bf2-kube-api-access-lhjcx\") pod \"openstackclient\" (UID: \"090560a3-12be-4a67-b507-0c7764053bf2\") " pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.178845 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data-custom\") pod \"983737cc-d82d-4d76-a060-b27cb8bcd94d\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.178881 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/983737cc-d82d-4d76-a060-b27cb8bcd94d-etc-machine-id\") pod \"983737cc-d82d-4d76-a060-b27cb8bcd94d\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.178938 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-combined-ca-bundle\") pod \"983737cc-d82d-4d76-a060-b27cb8bcd94d\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.179029 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data\") pod \"983737cc-d82d-4d76-a060-b27cb8bcd94d\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.179085 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn4bl\" (UniqueName: \"kubernetes.io/projected/983737cc-d82d-4d76-a060-b27cb8bcd94d-kube-api-access-rn4bl\") pod \"983737cc-d82d-4d76-a060-b27cb8bcd94d\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.179127 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-scripts\") pod \"983737cc-d82d-4d76-a060-b27cb8bcd94d\" (UID: \"983737cc-d82d-4d76-a060-b27cb8bcd94d\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.181458 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/983737cc-d82d-4d76-a060-b27cb8bcd94d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "983737cc-d82d-4d76-a060-b27cb8bcd94d" (UID: "983737cc-d82d-4d76-a060-b27cb8bcd94d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.184056 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "983737cc-d82d-4d76-a060-b27cb8bcd94d" (UID: "983737cc-d82d-4d76-a060-b27cb8bcd94d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.184422 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-scripts" (OuterVolumeSpecName: "scripts") pod "983737cc-d82d-4d76-a060-b27cb8bcd94d" (UID: "983737cc-d82d-4d76-a060-b27cb8bcd94d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.186468 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/983737cc-d82d-4d76-a060-b27cb8bcd94d-kube-api-access-rn4bl" (OuterVolumeSpecName: "kube-api-access-rn4bl") pod "983737cc-d82d-4d76-a060-b27cb8bcd94d" (UID: "983737cc-d82d-4d76-a060-b27cb8bcd94d"). InnerVolumeSpecName "kube-api-access-rn4bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.238431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "983737cc-d82d-4d76-a060-b27cb8bcd94d" (UID: "983737cc-d82d-4d76-a060-b27cb8bcd94d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.278446 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data" (OuterVolumeSpecName: "config-data") pod "983737cc-d82d-4d76-a060-b27cb8bcd94d" (UID: "983737cc-d82d-4d76-a060-b27cb8bcd94d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.281530 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.281572 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn4bl\" (UniqueName: \"kubernetes.io/projected/983737cc-d82d-4d76-a060-b27cb8bcd94d-kube-api-access-rn4bl\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.281588 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.281600 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.281612 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/983737cc-d82d-4d76-a060-b27cb8bcd94d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.281622 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983737cc-d82d-4d76-a060-b27cb8bcd94d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.306440 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.487647 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.489661 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.489762 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"983737cc-d82d-4d76-a060-b27cb8bcd94d","Type":"ContainerDied","Data":"299cdd87e424cf871fde98ea449c49ce26876a78710d023a8f7f3948527ea278"} Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.489837 4809 scope.go:117] "RemoveContainer" containerID="11e20e37adde989f6d2996800e67f6f3713a090171d816faf51ea99512f137bf" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.498225 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="79db49b5-8439-451b-936d-c4021dc7567e" podUID="090560a3-12be-4a67-b507-0c7764053bf2" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.557963 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.566400 4809 scope.go:117] "RemoveContainer" containerID="4361680b2d04b78af6a69f3c07134a57666e9e73d712670abf2504e350e9a498" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.596814 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.610441 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.617651 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 17:39:56 crc kubenswrapper[4809]: E1127 17:39:56.618168 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerName="probe" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.618189 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerName="probe" Nov 27 17:39:56 crc kubenswrapper[4809]: E1127 17:39:56.618211 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerName="cinder-scheduler" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.618217 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerName="cinder-scheduler" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.618381 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerName="cinder-scheduler" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.618409 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="983737cc-d82d-4d76-a060-b27cb8bcd94d" containerName="probe" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.619585 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.623510 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.625123 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.690437 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config\") pod \"79db49b5-8439-451b-936d-c4021dc7567e\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.690503 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9zwv\" (UniqueName: \"kubernetes.io/projected/79db49b5-8439-451b-936d-c4021dc7567e-kube-api-access-v9zwv\") pod \"79db49b5-8439-451b-936d-c4021dc7567e\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.690531 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-combined-ca-bundle\") pod \"79db49b5-8439-451b-936d-c4021dc7567e\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.690562 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config-secret\") pod \"79db49b5-8439-451b-936d-c4021dc7567e\" (UID: \"79db49b5-8439-451b-936d-c4021dc7567e\") " Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.690993 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-config-data\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.691039 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-scripts\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.691082 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzvml\" (UniqueName: \"kubernetes.io/projected/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-kube-api-access-kzvml\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.691107 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.691147 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.691203 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.692355 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "79db49b5-8439-451b-936d-c4021dc7567e" (UID: "79db49b5-8439-451b-936d-c4021dc7567e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.699418 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79db49b5-8439-451b-936d-c4021dc7567e-kube-api-access-v9zwv" (OuterVolumeSpecName: "kube-api-access-v9zwv") pod "79db49b5-8439-451b-936d-c4021dc7567e" (UID: "79db49b5-8439-451b-936d-c4021dc7567e"). InnerVolumeSpecName "kube-api-access-v9zwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.709883 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "79db49b5-8439-451b-936d-c4021dc7567e" (UID: "79db49b5-8439-451b-936d-c4021dc7567e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.718781 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79db49b5-8439-451b-936d-c4021dc7567e" (UID: "79db49b5-8439-451b-936d-c4021dc7567e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793095 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-config-data\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793152 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-scripts\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzvml\" (UniqueName: \"kubernetes.io/projected/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-kube-api-access-kzvml\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793242 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793290 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793387 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793399 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9zwv\" (UniqueName: \"kubernetes.io/projected/79db49b5-8439-451b-936d-c4021dc7567e-kube-api-access-v9zwv\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793409 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.793416 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79db49b5-8439-451b-936d-c4021dc7567e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.798349 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.799654 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.803535 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-scripts\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.803607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.803965 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.809120 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-config-data\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.820874 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzvml\" (UniqueName: \"kubernetes.io/projected/5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92-kube-api-access-kzvml\") pod \"cinder-scheduler-0\" (UID: \"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92\") " pod="openstack/cinder-scheduler-0" Nov 27 17:39:56 crc kubenswrapper[4809]: I1127 17:39:56.950055 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 17:39:57 crc kubenswrapper[4809]: I1127 17:39:57.421757 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 17:39:57 crc kubenswrapper[4809]: W1127 17:39:57.424878 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ea00edc_9b23_4bbe_93cc_cfae1ca2ff92.slice/crio-41a2ab7b4d97f4df9381daa947daebef684352c2d43a1c1f217620b030b8b210 WatchSource:0}: Error finding container 41a2ab7b4d97f4df9381daa947daebef684352c2d43a1c1f217620b030b8b210: Status 404 returned error can't find the container with id 41a2ab7b4d97f4df9381daa947daebef684352c2d43a1c1f217620b030b8b210 Nov 27 17:39:57 crc kubenswrapper[4809]: I1127 17:39:57.495699 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79db49b5-8439-451b-936d-c4021dc7567e" path="/var/lib/kubelet/pods/79db49b5-8439-451b-936d-c4021dc7567e/volumes" Nov 27 17:39:57 crc kubenswrapper[4809]: I1127 17:39:57.496483 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="983737cc-d82d-4d76-a060-b27cb8bcd94d" path="/var/lib/kubelet/pods/983737cc-d82d-4d76-a060-b27cb8bcd94d/volumes" Nov 27 17:39:57 crc kubenswrapper[4809]: I1127 17:39:57.507873 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"090560a3-12be-4a67-b507-0c7764053bf2","Type":"ContainerStarted","Data":"d511affa6add157e38ed52500507675488afbbc6820f6699621aa0511202722c"} Nov 27 17:39:57 crc kubenswrapper[4809]: I1127 17:39:57.509322 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 17:39:57 crc kubenswrapper[4809]: I1127 17:39:57.509619 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92","Type":"ContainerStarted","Data":"41a2ab7b4d97f4df9381daa947daebef684352c2d43a1c1f217620b030b8b210"} Nov 27 17:39:57 crc kubenswrapper[4809]: I1127 17:39:57.518762 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="79db49b5-8439-451b-936d-c4021dc7567e" podUID="090560a3-12be-4a67-b507-0c7764053bf2" Nov 27 17:39:58 crc kubenswrapper[4809]: I1127 17:39:58.545472 4809 generic.go:334] "Generic (PLEG): container finished" podID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerID="9035331522b1447565598f7473b7ab808d2ec9ca68cd09992c867d8b5192020b" exitCode=0 Nov 27 17:39:58 crc kubenswrapper[4809]: I1127 17:39:58.545979 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffcf7bfb-7p9tr" event={"ID":"013eaaea-f9de-4e5c-9647-a5df367b7709","Type":"ContainerDied","Data":"9035331522b1447565598f7473b7ab808d2ec9ca68cd09992c867d8b5192020b"} Nov 27 17:39:58 crc kubenswrapper[4809]: I1127 17:39:58.556060 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92","Type":"ContainerStarted","Data":"b1a50cff76901f58a42b368953f142418b0b7b52fd3d7ee179d90110daccbb7e"} Nov 27 17:39:58 crc kubenswrapper[4809]: I1127 17:39:58.912428 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.047412 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-combined-ca-bundle\") pod \"013eaaea-f9de-4e5c-9647-a5df367b7709\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.047480 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-ovndb-tls-certs\") pod \"013eaaea-f9de-4e5c-9647-a5df367b7709\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.047722 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-config\") pod \"013eaaea-f9de-4e5c-9647-a5df367b7709\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.047904 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnts9\" (UniqueName: \"kubernetes.io/projected/013eaaea-f9de-4e5c-9647-a5df367b7709-kube-api-access-cnts9\") pod \"013eaaea-f9de-4e5c-9647-a5df367b7709\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.047965 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-httpd-config\") pod \"013eaaea-f9de-4e5c-9647-a5df367b7709\" (UID: \"013eaaea-f9de-4e5c-9647-a5df367b7709\") " Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.066753 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "013eaaea-f9de-4e5c-9647-a5df367b7709" (UID: "013eaaea-f9de-4e5c-9647-a5df367b7709"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.069958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/013eaaea-f9de-4e5c-9647-a5df367b7709-kube-api-access-cnts9" (OuterVolumeSpecName: "kube-api-access-cnts9") pod "013eaaea-f9de-4e5c-9647-a5df367b7709" (UID: "013eaaea-f9de-4e5c-9647-a5df367b7709"). InnerVolumeSpecName "kube-api-access-cnts9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.124859 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-config" (OuterVolumeSpecName: "config") pod "013eaaea-f9de-4e5c-9647-a5df367b7709" (UID: "013eaaea-f9de-4e5c-9647-a5df367b7709"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.128607 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "013eaaea-f9de-4e5c-9647-a5df367b7709" (UID: "013eaaea-f9de-4e5c-9647-a5df367b7709"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.150521 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnts9\" (UniqueName: \"kubernetes.io/projected/013eaaea-f9de-4e5c-9647-a5df367b7709-kube-api-access-cnts9\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.150563 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.150575 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.150583 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.229550 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "013eaaea-f9de-4e5c-9647-a5df367b7709" (UID: "013eaaea-f9de-4e5c-9647-a5df367b7709"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.252062 4809 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/013eaaea-f9de-4e5c-9647-a5df367b7709-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.372319 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.573295 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffcf7bfb-7p9tr" event={"ID":"013eaaea-f9de-4e5c-9647-a5df367b7709","Type":"ContainerDied","Data":"798ad1c59a1972200e09171373a8932a578d98f2bfd8415b5e8ef0b97bddb5ae"} Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.573337 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cffcf7bfb-7p9tr" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.573370 4809 scope.go:117] "RemoveContainer" containerID="51d28b59af8a554b584b8f720373c960b56ded2c93ae0456729f8b49d97db81c" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.577969 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92","Type":"ContainerStarted","Data":"30e1366b3c88539eafb86cc2fca7018deb51199f26f664afdbf7e8b5c17ded7a"} Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.600403 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.600386023 podStartE2EDuration="3.600386023s" podCreationTimestamp="2025-11-27 17:39:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:39:59.592896479 +0000 UTC m=+1834.865353851" watchObservedRunningTime="2025-11-27 17:39:59.600386023 +0000 UTC m=+1834.872843375" Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.618902 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cffcf7bfb-7p9tr"] Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.635819 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6cffcf7bfb-7p9tr"] Nov 27 17:39:59 crc kubenswrapper[4809]: I1127 17:39:59.654631 4809 scope.go:117] "RemoveContainer" containerID="9035331522b1447565598f7473b7ab808d2ec9ca68cd09992c867d8b5192020b" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.529023 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5999bb96c5-jlmwd"] Nov 27 17:40:00 crc kubenswrapper[4809]: E1127 17:40:00.529913 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerName="neutron-httpd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.529938 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerName="neutron-httpd" Nov 27 17:40:00 crc kubenswrapper[4809]: E1127 17:40:00.529966 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerName="neutron-api" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.529975 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerName="neutron-api" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.530259 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerName="neutron-httpd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.530297 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="013eaaea-f9de-4e5c-9647-a5df367b7709" containerName="neutron-api" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.531579 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.534310 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.534386 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.534536 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.542564 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5999bb96c5-jlmwd"] Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.691461 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82t4m\" (UniqueName: \"kubernetes.io/projected/2cdfac16-f905-422d-a059-60d8e3238bcf-kube-api-access-82t4m\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.691520 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-public-tls-certs\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.691563 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cdfac16-f905-422d-a059-60d8e3238bcf-run-httpd\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.691625 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2cdfac16-f905-422d-a059-60d8e3238bcf-etc-swift\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.691671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cdfac16-f905-422d-a059-60d8e3238bcf-log-httpd\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.691907 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-config-data\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.691958 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-combined-ca-bundle\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.691983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-internal-tls-certs\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.794233 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-config-data\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.794297 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-combined-ca-bundle\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.794318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-internal-tls-certs\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.794403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82t4m\" (UniqueName: \"kubernetes.io/projected/2cdfac16-f905-422d-a059-60d8e3238bcf-kube-api-access-82t4m\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.794426 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-public-tls-certs\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.794459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cdfac16-f905-422d-a059-60d8e3238bcf-run-httpd\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.794502 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2cdfac16-f905-422d-a059-60d8e3238bcf-etc-swift\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.794532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cdfac16-f905-422d-a059-60d8e3238bcf-log-httpd\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.795063 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cdfac16-f905-422d-a059-60d8e3238bcf-run-httpd\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.795095 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cdfac16-f905-422d-a059-60d8e3238bcf-log-httpd\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.801310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-internal-tls-certs\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.801364 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-config-data\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.802146 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-public-tls-certs\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.802589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2cdfac16-f905-422d-a059-60d8e3238bcf-etc-swift\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.809098 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cdfac16-f905-422d-a059-60d8e3238bcf-combined-ca-bundle\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.818217 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82t4m\" (UniqueName: \"kubernetes.io/projected/2cdfac16-f905-422d-a059-60d8e3238bcf-kube-api-access-82t4m\") pod \"swift-proxy-5999bb96c5-jlmwd\" (UID: \"2cdfac16-f905-422d-a059-60d8e3238bcf\") " pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:00 crc kubenswrapper[4809]: I1127 17:40:00.862536 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.390179 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5999bb96c5-jlmwd"] Nov 27 17:40:01 crc kubenswrapper[4809]: W1127 17:40:01.403209 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2cdfac16_f905_422d_a059_60d8e3238bcf.slice/crio-3536b9b00f046d7d1433697737b0f0857442378285e2d0b1aa6c06b738acfd41 WatchSource:0}: Error finding container 3536b9b00f046d7d1433697737b0f0857442378285e2d0b1aa6c06b738acfd41: Status 404 returned error can't find the container with id 3536b9b00f046d7d1433697737b0f0857442378285e2d0b1aa6c06b738acfd41 Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.458615 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:40:01 crc kubenswrapper[4809]: E1127 17:40:01.459021 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.468475 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="013eaaea-f9de-4e5c-9647-a5df367b7709" path="/var/lib/kubelet/pods/013eaaea-f9de-4e5c-9647-a5df367b7709/volumes" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.501710 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-7b4rl"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.503204 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.526184 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-7b4rl"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.602710 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-gkh6w"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.603931 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.613214 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0435257-8045-4fbb-8cd4-1c2857ec48ae-operator-scripts\") pod \"nova-api-db-create-7b4rl\" (UID: \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\") " pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.613426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv8mz\" (UniqueName: \"kubernetes.io/projected/a0435257-8045-4fbb-8cd4-1c2857ec48ae-kube-api-access-cv8mz\") pod \"nova-api-db-create-7b4rl\" (UID: \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\") " pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.615811 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gkh6w"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.615847 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5999bb96c5-jlmwd" event={"ID":"2cdfac16-f905-422d-a059-60d8e3238bcf","Type":"ContainerStarted","Data":"3536b9b00f046d7d1433697737b0f0857442378285e2d0b1aa6c06b738acfd41"} Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.630100 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5c84-account-create-update-crb4h"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.631271 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.642092 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.681585 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5c84-account-create-update-crb4h"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.712855 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-r6wt2"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.714173 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.715287 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddhv6\" (UniqueName: \"kubernetes.io/projected/5fca3206-aa78-48d2-a774-12ab7e83e2eb-kube-api-access-ddhv6\") pod \"nova-cell0-db-create-gkh6w\" (UID: \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\") " pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.715326 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18acc87-ba67-4098-9181-c522e2e22f64-operator-scripts\") pod \"nova-api-5c84-account-create-update-crb4h\" (UID: \"f18acc87-ba67-4098-9181-c522e2e22f64\") " pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.715364 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv8mz\" (UniqueName: \"kubernetes.io/projected/a0435257-8045-4fbb-8cd4-1c2857ec48ae-kube-api-access-cv8mz\") pod \"nova-api-db-create-7b4rl\" (UID: \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\") " pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.715441 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fca3206-aa78-48d2-a774-12ab7e83e2eb-operator-scripts\") pod \"nova-cell0-db-create-gkh6w\" (UID: \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\") " pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.717970 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qzld\" (UniqueName: \"kubernetes.io/projected/f18acc87-ba67-4098-9181-c522e2e22f64-kube-api-access-9qzld\") pod \"nova-api-5c84-account-create-update-crb4h\" (UID: \"f18acc87-ba67-4098-9181-c522e2e22f64\") " pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.718251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0435257-8045-4fbb-8cd4-1c2857ec48ae-operator-scripts\") pod \"nova-api-db-create-7b4rl\" (UID: \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\") " pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.719422 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0435257-8045-4fbb-8cd4-1c2857ec48ae-operator-scripts\") pod \"nova-api-db-create-7b4rl\" (UID: \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\") " pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.732459 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-r6wt2"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.734371 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv8mz\" (UniqueName: \"kubernetes.io/projected/a0435257-8045-4fbb-8cd4-1c2857ec48ae-kube-api-access-cv8mz\") pod \"nova-api-db-create-7b4rl\" (UID: \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\") " pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.817660 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-965d-account-create-update-qdw2l"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.819725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qzld\" (UniqueName: \"kubernetes.io/projected/f18acc87-ba67-4098-9181-c522e2e22f64-kube-api-access-9qzld\") pod \"nova-api-5c84-account-create-update-crb4h\" (UID: \"f18acc87-ba67-4098-9181-c522e2e22f64\") " pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.819841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddhv6\" (UniqueName: \"kubernetes.io/projected/5fca3206-aa78-48d2-a774-12ab7e83e2eb-kube-api-access-ddhv6\") pod \"nova-cell0-db-create-gkh6w\" (UID: \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\") " pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.819870 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d484\" (UniqueName: \"kubernetes.io/projected/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-kube-api-access-2d484\") pod \"nova-cell1-db-create-r6wt2\" (UID: \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\") " pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.819893 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18acc87-ba67-4098-9181-c522e2e22f64-operator-scripts\") pod \"nova-api-5c84-account-create-update-crb4h\" (UID: \"f18acc87-ba67-4098-9181-c522e2e22f64\") " pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.819938 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fca3206-aa78-48d2-a774-12ab7e83e2eb-operator-scripts\") pod \"nova-cell0-db-create-gkh6w\" (UID: \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\") " pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.819964 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-operator-scripts\") pod \"nova-cell1-db-create-r6wt2\" (UID: \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\") " pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.820307 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.821468 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fca3206-aa78-48d2-a774-12ab7e83e2eb-operator-scripts\") pod \"nova-cell0-db-create-gkh6w\" (UID: \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\") " pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.821779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18acc87-ba67-4098-9181-c522e2e22f64-operator-scripts\") pod \"nova-api-5c84-account-create-update-crb4h\" (UID: \"f18acc87-ba67-4098-9181-c522e2e22f64\") " pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.824464 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.825103 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.839333 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-965d-account-create-update-qdw2l"] Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.848799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qzld\" (UniqueName: \"kubernetes.io/projected/f18acc87-ba67-4098-9181-c522e2e22f64-kube-api-access-9qzld\") pod \"nova-api-5c84-account-create-update-crb4h\" (UID: \"f18acc87-ba67-4098-9181-c522e2e22f64\") " pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.855367 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddhv6\" (UniqueName: \"kubernetes.io/projected/5fca3206-aa78-48d2-a774-12ab7e83e2eb-kube-api-access-ddhv6\") pod \"nova-cell0-db-create-gkh6w\" (UID: \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\") " pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.923403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d484\" (UniqueName: \"kubernetes.io/projected/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-kube-api-access-2d484\") pod \"nova-cell1-db-create-r6wt2\" (UID: \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\") " pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.923519 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-operator-scripts\") pod \"nova-cell1-db-create-r6wt2\" (UID: \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\") " pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.923620 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b01b2a4-9229-4e62-8606-44f8ef396563-operator-scripts\") pod \"nova-cell0-965d-account-create-update-qdw2l\" (UID: \"4b01b2a4-9229-4e62-8606-44f8ef396563\") " pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.923694 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4jxj\" (UniqueName: \"kubernetes.io/projected/4b01b2a4-9229-4e62-8606-44f8ef396563-kube-api-access-m4jxj\") pod \"nova-cell0-965d-account-create-update-qdw2l\" (UID: \"4b01b2a4-9229-4e62-8606-44f8ef396563\") " pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.924687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-operator-scripts\") pod \"nova-cell1-db-create-r6wt2\" (UID: \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\") " pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.943126 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d484\" (UniqueName: \"kubernetes.io/projected/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-kube-api-access-2d484\") pod \"nova-cell1-db-create-r6wt2\" (UID: \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\") " pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.950553 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.954368 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:01 crc kubenswrapper[4809]: I1127 17:40:01.971874 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.017927 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-846b-account-create-update-89v96"] Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.019384 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.021527 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.025996 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b01b2a4-9229-4e62-8606-44f8ef396563-operator-scripts\") pod \"nova-cell0-965d-account-create-update-qdw2l\" (UID: \"4b01b2a4-9229-4e62-8606-44f8ef396563\") " pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.026218 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4jxj\" (UniqueName: \"kubernetes.io/projected/4b01b2a4-9229-4e62-8606-44f8ef396563-kube-api-access-m4jxj\") pod \"nova-cell0-965d-account-create-update-qdw2l\" (UID: \"4b01b2a4-9229-4e62-8606-44f8ef396563\") " pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.029334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b01b2a4-9229-4e62-8606-44f8ef396563-operator-scripts\") pod \"nova-cell0-965d-account-create-update-qdw2l\" (UID: \"4b01b2a4-9229-4e62-8606-44f8ef396563\") " pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.031458 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.052778 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4jxj\" (UniqueName: \"kubernetes.io/projected/4b01b2a4-9229-4e62-8606-44f8ef396563-kube-api-access-m4jxj\") pod \"nova-cell0-965d-account-create-update-qdw2l\" (UID: \"4b01b2a4-9229-4e62-8606-44f8ef396563\") " pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.063548 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.064131 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-846b-account-create-update-89v96"] Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.128815 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/263114c5-20f2-45c1-b721-2b4d524da56f-operator-scripts\") pod \"nova-cell1-846b-account-create-update-89v96\" (UID: \"263114c5-20f2-45c1-b721-2b4d524da56f\") " pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.129169 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gcsf\" (UniqueName: \"kubernetes.io/projected/263114c5-20f2-45c1-b721-2b4d524da56f-kube-api-access-5gcsf\") pod \"nova-cell1-846b-account-create-update-89v96\" (UID: \"263114c5-20f2-45c1-b721-2b4d524da56f\") " pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.232619 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gcsf\" (UniqueName: \"kubernetes.io/projected/263114c5-20f2-45c1-b721-2b4d524da56f-kube-api-access-5gcsf\") pod \"nova-cell1-846b-account-create-update-89v96\" (UID: \"263114c5-20f2-45c1-b721-2b4d524da56f\") " pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.232686 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/263114c5-20f2-45c1-b721-2b4d524da56f-operator-scripts\") pod \"nova-cell1-846b-account-create-update-89v96\" (UID: \"263114c5-20f2-45c1-b721-2b4d524da56f\") " pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.233516 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/263114c5-20f2-45c1-b721-2b4d524da56f-operator-scripts\") pod \"nova-cell1-846b-account-create-update-89v96\" (UID: \"263114c5-20f2-45c1-b721-2b4d524da56f\") " pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.251779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gcsf\" (UniqueName: \"kubernetes.io/projected/263114c5-20f2-45c1-b721-2b4d524da56f-kube-api-access-5gcsf\") pod \"nova-cell1-846b-account-create-update-89v96\" (UID: \"263114c5-20f2-45c1-b721-2b4d524da56f\") " pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.351913 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-7b4rl"] Nov 27 17:40:02 crc kubenswrapper[4809]: W1127 17:40:02.359938 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0435257_8045_4fbb_8cd4_1c2857ec48ae.slice/crio-ef12d9c84214293cc577aa75afd76d9693959fbe1ee61c41bf656395b2839732 WatchSource:0}: Error finding container ef12d9c84214293cc577aa75afd76d9693959fbe1ee61c41bf656395b2839732: Status 404 returned error can't find the container with id ef12d9c84214293cc577aa75afd76d9693959fbe1ee61c41bf656395b2839732 Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.375372 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:02 crc kubenswrapper[4809]: W1127 17:40:02.536766 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fca3206_aa78_48d2_a774_12ab7e83e2eb.slice/crio-bd1698947e9700d584bae5c455725986ed56c620391ad9a3d5dd20c4a2ef8947 WatchSource:0}: Error finding container bd1698947e9700d584bae5c455725986ed56c620391ad9a3d5dd20c4a2ef8947: Status 404 returned error can't find the container with id bd1698947e9700d584bae5c455725986ed56c620391ad9a3d5dd20c4a2ef8947 Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.558192 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gkh6w"] Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.637049 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-7b4rl" event={"ID":"a0435257-8045-4fbb-8cd4-1c2857ec48ae","Type":"ContainerStarted","Data":"ef12d9c84214293cc577aa75afd76d9693959fbe1ee61c41bf656395b2839732"} Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.640444 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gkh6w" event={"ID":"5fca3206-aa78-48d2-a774-12ab7e83e2eb","Type":"ContainerStarted","Data":"bd1698947e9700d584bae5c455725986ed56c620391ad9a3d5dd20c4a2ef8947"} Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.652759 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5999bb96c5-jlmwd" event={"ID":"2cdfac16-f905-422d-a059-60d8e3238bcf","Type":"ContainerStarted","Data":"c572cd29efec567454256a6f8eb290862b034eb6ada5d0721568600db311cf33"} Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.653055 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.653066 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5999bb96c5-jlmwd" event={"ID":"2cdfac16-f905-422d-a059-60d8e3238bcf","Type":"ContainerStarted","Data":"c2c8029865d1a57bd1f919cd0ed113cb407fa472fde550ac2cf9c6dd1c888b9c"} Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.653304 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.880774 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5999bb96c5-jlmwd" podStartSLOduration=2.880731787 podStartE2EDuration="2.880731787s" podCreationTimestamp="2025-11-27 17:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:02.686512121 +0000 UTC m=+1837.958969493" watchObservedRunningTime="2025-11-27 17:40:02.880731787 +0000 UTC m=+1838.153189139" Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.884160 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-r6wt2"] Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.900041 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-965d-account-create-update-qdw2l"] Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.920259 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5c84-account-create-update-crb4h"] Nov 27 17:40:02 crc kubenswrapper[4809]: I1127 17:40:02.989068 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-846b-account-create-update-89v96"] Nov 27 17:40:03 crc kubenswrapper[4809]: W1127 17:40:03.017923 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod263114c5_20f2_45c1_b721_2b4d524da56f.slice/crio-f96fd4999ebb50624792d584cda847073cf81d46e095cc016115db9b26d18d54 WatchSource:0}: Error finding container f96fd4999ebb50624792d584cda847073cf81d46e095cc016115db9b26d18d54: Status 404 returned error can't find the container with id f96fd4999ebb50624792d584cda847073cf81d46e095cc016115db9b26d18d54 Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.707809 4809 generic.go:334] "Generic (PLEG): container finished" podID="0f2a7dfd-6adf-42c5-956d-4f148cea25b2" containerID="e0b4e8e3098c5cd99e7e226bf1fe4a931793eb0aa9fdfc9ddd33cb32a936029b" exitCode=0 Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.708226 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r6wt2" event={"ID":"0f2a7dfd-6adf-42c5-956d-4f148cea25b2","Type":"ContainerDied","Data":"e0b4e8e3098c5cd99e7e226bf1fe4a931793eb0aa9fdfc9ddd33cb32a936029b"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.708258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r6wt2" event={"ID":"0f2a7dfd-6adf-42c5-956d-4f148cea25b2","Type":"ContainerStarted","Data":"caba55a200f24561694582aed66687a35568ba5e76eed8926f42468ef5aaa8ee"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.711933 4809 generic.go:334] "Generic (PLEG): container finished" podID="4b01b2a4-9229-4e62-8606-44f8ef396563" containerID="9046c1399ae606399c4084ee74581c245bf9bbd1b1bf6aa167c182925ad22071" exitCode=0 Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.711998 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-965d-account-create-update-qdw2l" event={"ID":"4b01b2a4-9229-4e62-8606-44f8ef396563","Type":"ContainerDied","Data":"9046c1399ae606399c4084ee74581c245bf9bbd1b1bf6aa167c182925ad22071"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.713840 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-965d-account-create-update-qdw2l" event={"ID":"4b01b2a4-9229-4e62-8606-44f8ef396563","Type":"ContainerStarted","Data":"667ebade1d328ffa0fc2f52e76351d8e862d550b2265985ad23d1abb7593a11b"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.732936 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-846b-account-create-update-89v96" event={"ID":"263114c5-20f2-45c1-b721-2b4d524da56f","Type":"ContainerStarted","Data":"d25e17e852070998e7ac6280459ca4ae23bb3bbdb7b762032229e0d82294fb48"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.733013 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-846b-account-create-update-89v96" event={"ID":"263114c5-20f2-45c1-b721-2b4d524da56f","Type":"ContainerStarted","Data":"f96fd4999ebb50624792d584cda847073cf81d46e095cc016115db9b26d18d54"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.756232 4809 generic.go:334] "Generic (PLEG): container finished" podID="a0435257-8045-4fbb-8cd4-1c2857ec48ae" containerID="99d1afd96f55db6ed17b3d77809d95307eee5184ad099123dc966d13531c3818" exitCode=0 Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.756500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-7b4rl" event={"ID":"a0435257-8045-4fbb-8cd4-1c2857ec48ae","Type":"ContainerDied","Data":"99d1afd96f55db6ed17b3d77809d95307eee5184ad099123dc966d13531c3818"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.776877 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5c84-account-create-update-crb4h" event={"ID":"f18acc87-ba67-4098-9181-c522e2e22f64","Type":"ContainerStarted","Data":"ad609c13d8ab3eed8a920f1426b5844dd3c60ceffe004309963bbba33ab5ca66"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.776927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5c84-account-create-update-crb4h" event={"ID":"f18acc87-ba67-4098-9181-c522e2e22f64","Type":"ContainerStarted","Data":"af29f5ba103dbaaaa3e172efe01c671f3c445215188cb1332f127de19a54ca0d"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.788847 4809 generic.go:334] "Generic (PLEG): container finished" podID="5fca3206-aa78-48d2-a774-12ab7e83e2eb" containerID="d6ed7c84c17e5b83a62f584fc551e0c13e4780b476263cbbf35569276d45f7d3" exitCode=0 Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.790099 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gkh6w" event={"ID":"5fca3206-aa78-48d2-a774-12ab7e83e2eb","Type":"ContainerDied","Data":"d6ed7c84c17e5b83a62f584fc551e0c13e4780b476263cbbf35569276d45f7d3"} Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.799848 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-846b-account-create-update-89v96" podStartSLOduration=2.799820353 podStartE2EDuration="2.799820353s" podCreationTimestamp="2025-11-27 17:40:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:03.766468646 +0000 UTC m=+1839.038925998" watchObservedRunningTime="2025-11-27 17:40:03.799820353 +0000 UTC m=+1839.072277705" Nov 27 17:40:03 crc kubenswrapper[4809]: I1127 17:40:03.842945 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-5c84-account-create-update-crb4h" podStartSLOduration=2.842925507 podStartE2EDuration="2.842925507s" podCreationTimestamp="2025-11-27 17:40:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:03.818365428 +0000 UTC m=+1839.090822800" watchObservedRunningTime="2025-11-27 17:40:03.842925507 +0000 UTC m=+1839.115382859" Nov 27 17:40:04 crc kubenswrapper[4809]: I1127 17:40:04.803829 4809 generic.go:334] "Generic (PLEG): container finished" podID="263114c5-20f2-45c1-b721-2b4d524da56f" containerID="d25e17e852070998e7ac6280459ca4ae23bb3bbdb7b762032229e0d82294fb48" exitCode=0 Nov 27 17:40:04 crc kubenswrapper[4809]: I1127 17:40:04.803910 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-846b-account-create-update-89v96" event={"ID":"263114c5-20f2-45c1-b721-2b4d524da56f","Type":"ContainerDied","Data":"d25e17e852070998e7ac6280459ca4ae23bb3bbdb7b762032229e0d82294fb48"} Nov 27 17:40:04 crc kubenswrapper[4809]: I1127 17:40:04.806347 4809 generic.go:334] "Generic (PLEG): container finished" podID="596ec651-2088-4459-b0dd-224c67150b5d" containerID="d1ea97d252dae516cc60372d4b582e962b4cb543468583ae0dbf65f9ebf7504c" exitCode=137 Nov 27 17:40:04 crc kubenswrapper[4809]: I1127 17:40:04.806397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596ec651-2088-4459-b0dd-224c67150b5d","Type":"ContainerDied","Data":"d1ea97d252dae516cc60372d4b582e962b4cb543468583ae0dbf65f9ebf7504c"} Nov 27 17:40:04 crc kubenswrapper[4809]: I1127 17:40:04.808135 4809 generic.go:334] "Generic (PLEG): container finished" podID="f18acc87-ba67-4098-9181-c522e2e22f64" containerID="ad609c13d8ab3eed8a920f1426b5844dd3c60ceffe004309963bbba33ab5ca66" exitCode=0 Nov 27 17:40:04 crc kubenswrapper[4809]: I1127 17:40:04.808335 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5c84-account-create-update-crb4h" event={"ID":"f18acc87-ba67-4098-9181-c522e2e22f64","Type":"ContainerDied","Data":"ad609c13d8ab3eed8a920f1426b5844dd3c60ceffe004309963bbba33ab5ca66"} Nov 27 17:40:05 crc kubenswrapper[4809]: I1127 17:40:05.598349 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:40:05 crc kubenswrapper[4809]: I1127 17:40:05.598932 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerName="glance-log" containerID="cri-o://174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045" gracePeriod=30 Nov 27 17:40:05 crc kubenswrapper[4809]: I1127 17:40:05.599177 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerName="glance-httpd" containerID="cri-o://fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97" gracePeriod=30 Nov 27 17:40:05 crc kubenswrapper[4809]: I1127 17:40:05.608561 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/glance-default-internal-api-0" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.165:9292/healthcheck\": EOF" Nov 27 17:40:05 crc kubenswrapper[4809]: I1127 17:40:05.830522 4809 generic.go:334] "Generic (PLEG): container finished" podID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerID="174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045" exitCode=143 Nov 27 17:40:05 crc kubenswrapper[4809]: I1127 17:40:05.830722 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7f5d5190-5f50-4d4e-9739-250c85c0146b","Type":"ContainerDied","Data":"174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045"} Nov 27 17:40:06 crc kubenswrapper[4809]: I1127 17:40:06.513143 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:40:06 crc kubenswrapper[4809]: I1127 17:40:06.513487 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-log" containerID="cri-o://4c303b3aa29b3f559a9eedeb7784e5a607de6c46cda79f3c034b0422b71cbd76" gracePeriod=30 Nov 27 17:40:06 crc kubenswrapper[4809]: I1127 17:40:06.513571 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-httpd" containerID="cri-o://6bd864cab97711e66dabfdb917ca4f1b88372afed903288f05688afb3b95b68f" gracePeriod=30 Nov 27 17:40:06 crc kubenswrapper[4809]: I1127 17:40:06.844583 4809 generic.go:334] "Generic (PLEG): container finished" podID="e345d6af-f198-42b0-9017-582c93ffeef6" containerID="4c303b3aa29b3f559a9eedeb7784e5a607de6c46cda79f3c034b0422b71cbd76" exitCode=143 Nov 27 17:40:06 crc kubenswrapper[4809]: I1127 17:40:06.844669 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e345d6af-f198-42b0-9017-582c93ffeef6","Type":"ContainerDied","Data":"4c303b3aa29b3f559a9eedeb7784e5a607de6c46cda79f3c034b0422b71cbd76"} Nov 27 17:40:07 crc kubenswrapper[4809]: I1127 17:40:07.165124 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 27 17:40:08 crc kubenswrapper[4809]: I1127 17:40:08.870312 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-965d-account-create-update-qdw2l" event={"ID":"4b01b2a4-9229-4e62-8606-44f8ef396563","Type":"ContainerDied","Data":"667ebade1d328ffa0fc2f52e76351d8e862d550b2265985ad23d1abb7593a11b"} Nov 27 17:40:08 crc kubenswrapper[4809]: I1127 17:40:08.870725 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="667ebade1d328ffa0fc2f52e76351d8e862d550b2265985ad23d1abb7593a11b" Nov 27 17:40:08 crc kubenswrapper[4809]: I1127 17:40:08.872198 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-7b4rl" event={"ID":"a0435257-8045-4fbb-8cd4-1c2857ec48ae","Type":"ContainerDied","Data":"ef12d9c84214293cc577aa75afd76d9693959fbe1ee61c41bf656395b2839732"} Nov 27 17:40:08 crc kubenswrapper[4809]: I1127 17:40:08.872220 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef12d9c84214293cc577aa75afd76d9693959fbe1ee61c41bf656395b2839732" Nov 27 17:40:08 crc kubenswrapper[4809]: I1127 17:40:08.874011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gkh6w" event={"ID":"5fca3206-aa78-48d2-a774-12ab7e83e2eb","Type":"ContainerDied","Data":"bd1698947e9700d584bae5c455725986ed56c620391ad9a3d5dd20c4a2ef8947"} Nov 27 17:40:08 crc kubenswrapper[4809]: I1127 17:40:08.874036 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd1698947e9700d584bae5c455725986ed56c620391ad9a3d5dd20c4a2ef8947" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.161687 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.172999 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.222884 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.234258 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.250941 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.258751 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.281898 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312369 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0435257-8045-4fbb-8cd4-1c2857ec48ae-operator-scripts\") pod \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\" (UID: \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312486 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddhv6\" (UniqueName: \"kubernetes.io/projected/5fca3206-aa78-48d2-a774-12ab7e83e2eb-kube-api-access-ddhv6\") pod \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\" (UID: \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312574 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d484\" (UniqueName: \"kubernetes.io/projected/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-kube-api-access-2d484\") pod \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\" (UID: \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312591 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4jxj\" (UniqueName: \"kubernetes.io/projected/4b01b2a4-9229-4e62-8606-44f8ef396563-kube-api-access-m4jxj\") pod \"4b01b2a4-9229-4e62-8606-44f8ef396563\" (UID: \"4b01b2a4-9229-4e62-8606-44f8ef396563\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312677 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv8mz\" (UniqueName: \"kubernetes.io/projected/a0435257-8045-4fbb-8cd4-1c2857ec48ae-kube-api-access-cv8mz\") pod \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\" (UID: \"a0435257-8045-4fbb-8cd4-1c2857ec48ae\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312697 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fca3206-aa78-48d2-a774-12ab7e83e2eb-operator-scripts\") pod \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\" (UID: \"5fca3206-aa78-48d2-a774-12ab7e83e2eb\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-operator-scripts\") pod \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\" (UID: \"0f2a7dfd-6adf-42c5-956d-4f148cea25b2\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312754 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b01b2a4-9229-4e62-8606-44f8ef396563-operator-scripts\") pod \"4b01b2a4-9229-4e62-8606-44f8ef396563\" (UID: \"4b01b2a4-9229-4e62-8606-44f8ef396563\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/263114c5-20f2-45c1-b721-2b4d524da56f-operator-scripts\") pod \"263114c5-20f2-45c1-b721-2b4d524da56f\" (UID: \"263114c5-20f2-45c1-b721-2b4d524da56f\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312848 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18acc87-ba67-4098-9181-c522e2e22f64-operator-scripts\") pod \"f18acc87-ba67-4098-9181-c522e2e22f64\" (UID: \"f18acc87-ba67-4098-9181-c522e2e22f64\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312879 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qzld\" (UniqueName: \"kubernetes.io/projected/f18acc87-ba67-4098-9181-c522e2e22f64-kube-api-access-9qzld\") pod \"f18acc87-ba67-4098-9181-c522e2e22f64\" (UID: \"f18acc87-ba67-4098-9181-c522e2e22f64\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.312912 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gcsf\" (UniqueName: \"kubernetes.io/projected/263114c5-20f2-45c1-b721-2b4d524da56f-kube-api-access-5gcsf\") pod \"263114c5-20f2-45c1-b721-2b4d524da56f\" (UID: \"263114c5-20f2-45c1-b721-2b4d524da56f\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.313410 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f2a7dfd-6adf-42c5-956d-4f148cea25b2" (UID: "0f2a7dfd-6adf-42c5-956d-4f148cea25b2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.313448 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0435257-8045-4fbb-8cd4-1c2857ec48ae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a0435257-8045-4fbb-8cd4-1c2857ec48ae" (UID: "a0435257-8045-4fbb-8cd4-1c2857ec48ae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.314759 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/263114c5-20f2-45c1-b721-2b4d524da56f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "263114c5-20f2-45c1-b721-2b4d524da56f" (UID: "263114c5-20f2-45c1-b721-2b4d524da56f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.315193 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b01b2a4-9229-4e62-8606-44f8ef396563-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4b01b2a4-9229-4e62-8606-44f8ef396563" (UID: "4b01b2a4-9229-4e62-8606-44f8ef396563"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.315603 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fca3206-aa78-48d2-a774-12ab7e83e2eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5fca3206-aa78-48d2-a774-12ab7e83e2eb" (UID: "5fca3206-aa78-48d2-a774-12ab7e83e2eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.322917 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f18acc87-ba67-4098-9181-c522e2e22f64-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f18acc87-ba67-4098-9181-c522e2e22f64" (UID: "f18acc87-ba67-4098-9181-c522e2e22f64"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.324709 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-kube-api-access-2d484" (OuterVolumeSpecName: "kube-api-access-2d484") pod "0f2a7dfd-6adf-42c5-956d-4f148cea25b2" (UID: "0f2a7dfd-6adf-42c5-956d-4f148cea25b2"). InnerVolumeSpecName "kube-api-access-2d484". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.324783 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/263114c5-20f2-45c1-b721-2b4d524da56f-kube-api-access-5gcsf" (OuterVolumeSpecName: "kube-api-access-5gcsf") pod "263114c5-20f2-45c1-b721-2b4d524da56f" (UID: "263114c5-20f2-45c1-b721-2b4d524da56f"). InnerVolumeSpecName "kube-api-access-5gcsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.325294 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fca3206-aa78-48d2-a774-12ab7e83e2eb-kube-api-access-ddhv6" (OuterVolumeSpecName: "kube-api-access-ddhv6") pod "5fca3206-aa78-48d2-a774-12ab7e83e2eb" (UID: "5fca3206-aa78-48d2-a774-12ab7e83e2eb"). InnerVolumeSpecName "kube-api-access-ddhv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.325906 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b01b2a4-9229-4e62-8606-44f8ef396563-kube-api-access-m4jxj" (OuterVolumeSpecName: "kube-api-access-m4jxj") pod "4b01b2a4-9229-4e62-8606-44f8ef396563" (UID: "4b01b2a4-9229-4e62-8606-44f8ef396563"). InnerVolumeSpecName "kube-api-access-m4jxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.328031 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0435257-8045-4fbb-8cd4-1c2857ec48ae-kube-api-access-cv8mz" (OuterVolumeSpecName: "kube-api-access-cv8mz") pod "a0435257-8045-4fbb-8cd4-1c2857ec48ae" (UID: "a0435257-8045-4fbb-8cd4-1c2857ec48ae"). InnerVolumeSpecName "kube-api-access-cv8mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.332215 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f18acc87-ba67-4098-9181-c522e2e22f64-kube-api-access-9qzld" (OuterVolumeSpecName: "kube-api-access-9qzld") pod "f18acc87-ba67-4098-9181-c522e2e22f64" (UID: "f18acc87-ba67-4098-9181-c522e2e22f64"). InnerVolumeSpecName "kube-api-access-9qzld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.414992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vlg4\" (UniqueName: \"kubernetes.io/projected/596ec651-2088-4459-b0dd-224c67150b5d-kube-api-access-2vlg4\") pod \"596ec651-2088-4459-b0dd-224c67150b5d\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.417201 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-config-data\") pod \"596ec651-2088-4459-b0dd-224c67150b5d\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.417840 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-combined-ca-bundle\") pod \"596ec651-2088-4459-b0dd-224c67150b5d\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.418182 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-scripts\") pod \"596ec651-2088-4459-b0dd-224c67150b5d\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.418293 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-log-httpd\") pod \"596ec651-2088-4459-b0dd-224c67150b5d\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.418632 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-sg-core-conf-yaml\") pod \"596ec651-2088-4459-b0dd-224c67150b5d\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.418784 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/596ec651-2088-4459-b0dd-224c67150b5d-kube-api-access-2vlg4" (OuterVolumeSpecName: "kube-api-access-2vlg4") pod "596ec651-2088-4459-b0dd-224c67150b5d" (UID: "596ec651-2088-4459-b0dd-224c67150b5d"). InnerVolumeSpecName "kube-api-access-2vlg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.418809 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-run-httpd\") pod \"596ec651-2088-4459-b0dd-224c67150b5d\" (UID: \"596ec651-2088-4459-b0dd-224c67150b5d\") " Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.418845 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "596ec651-2088-4459-b0dd-224c67150b5d" (UID: "596ec651-2088-4459-b0dd-224c67150b5d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.419496 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "596ec651-2088-4459-b0dd-224c67150b5d" (UID: "596ec651-2088-4459-b0dd-224c67150b5d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420168 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0435257-8045-4fbb-8cd4-1c2857ec48ae-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420218 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddhv6\" (UniqueName: \"kubernetes.io/projected/5fca3206-aa78-48d2-a774-12ab7e83e2eb-kube-api-access-ddhv6\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420231 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d484\" (UniqueName: \"kubernetes.io/projected/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-kube-api-access-2d484\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420244 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4jxj\" (UniqueName: \"kubernetes.io/projected/4b01b2a4-9229-4e62-8606-44f8ef396563-kube-api-access-m4jxj\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420255 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420290 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/596ec651-2088-4459-b0dd-224c67150b5d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420302 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv8mz\" (UniqueName: \"kubernetes.io/projected/a0435257-8045-4fbb-8cd4-1c2857ec48ae-kube-api-access-cv8mz\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420313 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fca3206-aa78-48d2-a774-12ab7e83e2eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420325 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f2a7dfd-6adf-42c5-956d-4f148cea25b2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420335 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b01b2a4-9229-4e62-8606-44f8ef396563-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420367 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vlg4\" (UniqueName: \"kubernetes.io/projected/596ec651-2088-4459-b0dd-224c67150b5d-kube-api-access-2vlg4\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420378 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/263114c5-20f2-45c1-b721-2b4d524da56f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420389 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f18acc87-ba67-4098-9181-c522e2e22f64-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420904 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qzld\" (UniqueName: \"kubernetes.io/projected/f18acc87-ba67-4098-9181-c522e2e22f64-kube-api-access-9qzld\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.420952 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gcsf\" (UniqueName: \"kubernetes.io/projected/263114c5-20f2-45c1-b721-2b4d524da56f-kube-api-access-5gcsf\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.422145 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "596ec651-2088-4459-b0dd-224c67150b5d" (UID: "596ec651-2088-4459-b0dd-224c67150b5d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.422615 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-scripts" (OuterVolumeSpecName: "scripts") pod "596ec651-2088-4459-b0dd-224c67150b5d" (UID: "596ec651-2088-4459-b0dd-224c67150b5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.501610 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "596ec651-2088-4459-b0dd-224c67150b5d" (UID: "596ec651-2088-4459-b0dd-224c67150b5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.523371 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.523400 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.523409 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.525303 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-config-data" (OuterVolumeSpecName: "config-data") pod "596ec651-2088-4459-b0dd-224c67150b5d" (UID: "596ec651-2088-4459-b0dd-224c67150b5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.625116 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/596ec651-2088-4459-b0dd-224c67150b5d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.660453 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.166:9292/healthcheck\": read tcp 10.217.0.2:34690->10.217.0.166:9292: read: connection reset by peer" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.660481 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.166:9292/healthcheck\": read tcp 10.217.0.2:34684->10.217.0.166:9292: read: connection reset by peer" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.887566 4809 generic.go:334] "Generic (PLEG): container finished" podID="e345d6af-f198-42b0-9017-582c93ffeef6" containerID="6bd864cab97711e66dabfdb917ca4f1b88372afed903288f05688afb3b95b68f" exitCode=0 Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.887659 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e345d6af-f198-42b0-9017-582c93ffeef6","Type":"ContainerDied","Data":"6bd864cab97711e66dabfdb917ca4f1b88372afed903288f05688afb3b95b68f"} Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.890339 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-846b-account-create-update-89v96" event={"ID":"263114c5-20f2-45c1-b721-2b4d524da56f","Type":"ContainerDied","Data":"f96fd4999ebb50624792d584cda847073cf81d46e095cc016115db9b26d18d54"} Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.890368 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f96fd4999ebb50624792d584cda847073cf81d46e095cc016115db9b26d18d54" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.890417 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-846b-account-create-update-89v96" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.894577 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"596ec651-2088-4459-b0dd-224c67150b5d","Type":"ContainerDied","Data":"587b6cf5aaed97ed37e874602721ffa02025c9d285cf1a12d87b4b1af6fd1ac1"} Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.894617 4809 scope.go:117] "RemoveContainer" containerID="d1ea97d252dae516cc60372d4b582e962b4cb543468583ae0dbf65f9ebf7504c" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.894793 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.905973 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"090560a3-12be-4a67-b507-0c7764053bf2","Type":"ContainerStarted","Data":"a4cad92f3d6668a1edbef96c512bb5c1820b0f9475dff407aa1db5dbffd5625b"} Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.908608 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5c84-account-create-update-crb4h" event={"ID":"f18acc87-ba67-4098-9181-c522e2e22f64","Type":"ContainerDied","Data":"af29f5ba103dbaaaa3e172efe01c671f3c445215188cb1332f127de19a54ca0d"} Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.908671 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af29f5ba103dbaaaa3e172efe01c671f3c445215188cb1332f127de19a54ca0d" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.908797 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5c84-account-create-update-crb4h" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.912565 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gkh6w" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.913464 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r6wt2" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.913710 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-965d-account-create-update-qdw2l" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.913722 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r6wt2" event={"ID":"0f2a7dfd-6adf-42c5-956d-4f148cea25b2","Type":"ContainerDied","Data":"caba55a200f24561694582aed66687a35568ba5e76eed8926f42468ef5aaa8ee"} Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.913775 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caba55a200f24561694582aed66687a35568ba5e76eed8926f42468ef5aaa8ee" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.913835 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7b4rl" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.932029 4809 scope.go:117] "RemoveContainer" containerID="ca99702884a990489a3335cd2835a1eec23536939f25917052640d89c57c60ae" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.935896 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.795315543 podStartE2EDuration="14.935869815s" podCreationTimestamp="2025-11-27 17:39:55 +0000 UTC" firstStartedPulling="2025-11-27 17:39:56.811605258 +0000 UTC m=+1832.084062610" lastFinishedPulling="2025-11-27 17:40:08.95215953 +0000 UTC m=+1844.224616882" observedRunningTime="2025-11-27 17:40:09.927509827 +0000 UTC m=+1845.199967169" watchObservedRunningTime="2025-11-27 17:40:09.935869815 +0000 UTC m=+1845.208327177" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.985963 4809 scope.go:117] "RemoveContainer" containerID="8f46e7899255c015a731ca5d6df54e769409b95cc0ab6ec66b6c162dea49ca74" Nov 27 17:40:09 crc kubenswrapper[4809]: I1127 17:40:09.995020 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.006261 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.017416 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:10 crc kubenswrapper[4809]: E1127 17:40:10.017965 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="ceilometer-central-agent" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.017983 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="ceilometer-central-agent" Nov 27 17:40:10 crc kubenswrapper[4809]: E1127 17:40:10.017996 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b01b2a4-9229-4e62-8606-44f8ef396563" containerName="mariadb-account-create-update" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018013 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b01b2a4-9229-4e62-8606-44f8ef396563" containerName="mariadb-account-create-update" Nov 27 17:40:10 crc kubenswrapper[4809]: E1127 17:40:10.018030 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fca3206-aa78-48d2-a774-12ab7e83e2eb" containerName="mariadb-database-create" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018040 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fca3206-aa78-48d2-a774-12ab7e83e2eb" containerName="mariadb-database-create" Nov 27 17:40:10 crc kubenswrapper[4809]: E1127 17:40:10.018061 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="proxy-httpd" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018070 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="proxy-httpd" Nov 27 17:40:10 crc kubenswrapper[4809]: E1127 17:40:10.018091 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="263114c5-20f2-45c1-b721-2b4d524da56f" containerName="mariadb-account-create-update" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018101 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="263114c5-20f2-45c1-b721-2b4d524da56f" containerName="mariadb-account-create-update" Nov 27 17:40:10 crc kubenswrapper[4809]: E1127 17:40:10.018117 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2a7dfd-6adf-42c5-956d-4f148cea25b2" containerName="mariadb-database-create" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018125 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2a7dfd-6adf-42c5-956d-4f148cea25b2" containerName="mariadb-database-create" Nov 27 17:40:10 crc kubenswrapper[4809]: E1127 17:40:10.018133 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18acc87-ba67-4098-9181-c522e2e22f64" containerName="mariadb-account-create-update" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018140 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18acc87-ba67-4098-9181-c522e2e22f64" containerName="mariadb-account-create-update" Nov 27 17:40:10 crc kubenswrapper[4809]: E1127 17:40:10.018155 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0435257-8045-4fbb-8cd4-1c2857ec48ae" containerName="mariadb-database-create" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018162 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0435257-8045-4fbb-8cd4-1c2857ec48ae" containerName="mariadb-database-create" Nov 27 17:40:10 crc kubenswrapper[4809]: E1127 17:40:10.018174 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="ceilometer-notification-agent" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018182 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="ceilometer-notification-agent" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018413 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="ceilometer-central-agent" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018435 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2a7dfd-6adf-42c5-956d-4f148cea25b2" containerName="mariadb-database-create" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018446 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0435257-8045-4fbb-8cd4-1c2857ec48ae" containerName="mariadb-database-create" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018459 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="263114c5-20f2-45c1-b721-2b4d524da56f" containerName="mariadb-account-create-update" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018477 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="proxy-httpd" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018495 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b01b2a4-9229-4e62-8606-44f8ef396563" containerName="mariadb-account-create-update" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018509 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f18acc87-ba67-4098-9181-c522e2e22f64" containerName="mariadb-account-create-update" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018522 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fca3206-aa78-48d2-a774-12ab7e83e2eb" containerName="mariadb-database-create" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.018534 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="596ec651-2088-4459-b0dd-224c67150b5d" containerName="ceilometer-notification-agent" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.020675 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.023448 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.023599 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.030937 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.137987 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwnpx\" (UniqueName: \"kubernetes.io/projected/3878cd07-3090-409e-add5-6d020076b63b-kube-api-access-mwnpx\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.138677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-log-httpd\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.138719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-scripts\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.138865 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.138972 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-config-data\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.139021 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.139114 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-run-httpd\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.258082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-config-data\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.258152 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.258205 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-run-httpd\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.258274 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwnpx\" (UniqueName: \"kubernetes.io/projected/3878cd07-3090-409e-add5-6d020076b63b-kube-api-access-mwnpx\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.258317 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-log-httpd\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.258344 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-scripts\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.258439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.259931 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-run-httpd\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.264489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-log-httpd\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.271180 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.273434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-scripts\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.275468 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.303348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-config-data\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.315079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwnpx\" (UniqueName: \"kubernetes.io/projected/3878cd07-3090-409e-add5-6d020076b63b-kube-api-access-mwnpx\") pod \"ceilometer-0\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.413553 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.414111 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.463015 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwt68\" (UniqueName: \"kubernetes.io/projected/e345d6af-f198-42b0-9017-582c93ffeef6-kube-api-access-fwt68\") pod \"e345d6af-f198-42b0-9017-582c93ffeef6\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.463131 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-combined-ca-bundle\") pod \"e345d6af-f198-42b0-9017-582c93ffeef6\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.463167 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-config-data\") pod \"e345d6af-f198-42b0-9017-582c93ffeef6\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.463294 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-logs\") pod \"e345d6af-f198-42b0-9017-582c93ffeef6\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.463375 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-scripts\") pod \"e345d6af-f198-42b0-9017-582c93ffeef6\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.463603 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"e345d6af-f198-42b0-9017-582c93ffeef6\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.463633 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-httpd-run\") pod \"e345d6af-f198-42b0-9017-582c93ffeef6\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.463733 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-public-tls-certs\") pod \"e345d6af-f198-42b0-9017-582c93ffeef6\" (UID: \"e345d6af-f198-42b0-9017-582c93ffeef6\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.464697 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-logs" (OuterVolumeSpecName: "logs") pod "e345d6af-f198-42b0-9017-582c93ffeef6" (UID: "e345d6af-f198-42b0-9017-582c93ffeef6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.465646 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e345d6af-f198-42b0-9017-582c93ffeef6" (UID: "e345d6af-f198-42b0-9017-582c93ffeef6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.469607 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e345d6af-f198-42b0-9017-582c93ffeef6-kube-api-access-fwt68" (OuterVolumeSpecName: "kube-api-access-fwt68") pod "e345d6af-f198-42b0-9017-582c93ffeef6" (UID: "e345d6af-f198-42b0-9017-582c93ffeef6"). InnerVolumeSpecName "kube-api-access-fwt68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.471856 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "e345d6af-f198-42b0-9017-582c93ffeef6" (UID: "e345d6af-f198-42b0-9017-582c93ffeef6"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.472211 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-scripts" (OuterVolumeSpecName: "scripts") pod "e345d6af-f198-42b0-9017-582c93ffeef6" (UID: "e345d6af-f198-42b0-9017-582c93ffeef6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.519931 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e345d6af-f198-42b0-9017-582c93ffeef6" (UID: "e345d6af-f198-42b0-9017-582c93ffeef6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.550464 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e345d6af-f198-42b0-9017-582c93ffeef6" (UID: "e345d6af-f198-42b0-9017-582c93ffeef6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.566802 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.566853 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.566868 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.566880 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.566894 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwt68\" (UniqueName: \"kubernetes.io/projected/e345d6af-f198-42b0-9017-582c93ffeef6-kube-api-access-fwt68\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.566905 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.566924 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e345d6af-f198-42b0-9017-582c93ffeef6-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.574540 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-config-data" (OuterVolumeSpecName: "config-data") pod "e345d6af-f198-42b0-9017-582c93ffeef6" (UID: "e345d6af-f198-42b0-9017-582c93ffeef6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.598085 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.672230 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.672266 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e345d6af-f198-42b0-9017-582c93ffeef6-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.689256 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.703968 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.773484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-httpd-run\") pod \"7f5d5190-5f50-4d4e-9739-250c85c0146b\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.773519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-scripts\") pod \"7f5d5190-5f50-4d4e-9739-250c85c0146b\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.773584 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-internal-tls-certs\") pod \"7f5d5190-5f50-4d4e-9739-250c85c0146b\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.773641 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gv524\" (UniqueName: \"kubernetes.io/projected/7f5d5190-5f50-4d4e-9739-250c85c0146b-kube-api-access-gv524\") pod \"7f5d5190-5f50-4d4e-9739-250c85c0146b\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.773680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"7f5d5190-5f50-4d4e-9739-250c85c0146b\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.773715 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-combined-ca-bundle\") pod \"7f5d5190-5f50-4d4e-9739-250c85c0146b\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.773776 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-logs\") pod \"7f5d5190-5f50-4d4e-9739-250c85c0146b\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.773858 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-config-data\") pod \"7f5d5190-5f50-4d4e-9739-250c85c0146b\" (UID: \"7f5d5190-5f50-4d4e-9739-250c85c0146b\") " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.774588 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7f5d5190-5f50-4d4e-9739-250c85c0146b" (UID: "7f5d5190-5f50-4d4e-9739-250c85c0146b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.775324 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-logs" (OuterVolumeSpecName: "logs") pod "7f5d5190-5f50-4d4e-9739-250c85c0146b" (UID: "7f5d5190-5f50-4d4e-9739-250c85c0146b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.778349 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-scripts" (OuterVolumeSpecName: "scripts") pod "7f5d5190-5f50-4d4e-9739-250c85c0146b" (UID: "7f5d5190-5f50-4d4e-9739-250c85c0146b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.779094 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f5d5190-5f50-4d4e-9739-250c85c0146b-kube-api-access-gv524" (OuterVolumeSpecName: "kube-api-access-gv524") pod "7f5d5190-5f50-4d4e-9739-250c85c0146b" (UID: "7f5d5190-5f50-4d4e-9739-250c85c0146b"). InnerVolumeSpecName "kube-api-access-gv524". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.792244 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "7f5d5190-5f50-4d4e-9739-250c85c0146b" (UID: "7f5d5190-5f50-4d4e-9739-250c85c0146b"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.822956 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f5d5190-5f50-4d4e-9739-250c85c0146b" (UID: "7f5d5190-5f50-4d4e-9739-250c85c0146b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.846493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7f5d5190-5f50-4d4e-9739-250c85c0146b" (UID: "7f5d5190-5f50-4d4e-9739-250c85c0146b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.848030 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-config-data" (OuterVolumeSpecName: "config-data") pod "7f5d5190-5f50-4d4e-9739-250c85c0146b" (UID: "7f5d5190-5f50-4d4e-9739-250c85c0146b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.876978 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.877762 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5999bb96c5-jlmwd" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.878879 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.878916 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.878927 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.878937 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.878947 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gv524\" (UniqueName: \"kubernetes.io/projected/7f5d5190-5f50-4d4e-9739-250c85c0146b-kube-api-access-gv524\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.879006 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.879020 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f5d5190-5f50-4d4e-9739-250c85c0146b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.879027 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f5d5190-5f50-4d4e-9739-250c85c0146b-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.906920 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.931024 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.931339 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e345d6af-f198-42b0-9017-582c93ffeef6","Type":"ContainerDied","Data":"2d9013182302b4f4ae1bb9bdc3942283fe488e5430143ea4669a08d9de4f7b3d"} Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.931384 4809 scope.go:117] "RemoveContainer" containerID="6bd864cab97711e66dabfdb917ca4f1b88372afed903288f05688afb3b95b68f" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.960297 4809 generic.go:334] "Generic (PLEG): container finished" podID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerID="fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97" exitCode=0 Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.960642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7f5d5190-5f50-4d4e-9739-250c85c0146b","Type":"ContainerDied","Data":"fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97"} Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.960698 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7f5d5190-5f50-4d4e-9739-250c85c0146b","Type":"ContainerDied","Data":"da35c38e53fdbb5e04f7d55ebf1badf31e3dbb8a47bd115f484fd92f45321e20"} Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.960799 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:10 crc kubenswrapper[4809]: I1127 17:40:10.983711 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.018834 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.027088 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.035052 4809 scope.go:117] "RemoveContainer" containerID="4c303b3aa29b3f559a9eedeb7784e5a607de6c46cda79f3c034b0422b71cbd76" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.036148 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.071861 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.099361 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.101082 4809 scope.go:117] "RemoveContainer" containerID="fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.111055 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:40:11 crc kubenswrapper[4809]: E1127 17:40:11.111532 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerName="glance-log" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.111548 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerName="glance-log" Nov 27 17:40:11 crc kubenswrapper[4809]: E1127 17:40:11.111576 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerName="glance-httpd" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.111584 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerName="glance-httpd" Nov 27 17:40:11 crc kubenswrapper[4809]: E1127 17:40:11.111605 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-httpd" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.111611 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-httpd" Nov 27 17:40:11 crc kubenswrapper[4809]: E1127 17:40:11.111622 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-log" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.111628 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-log" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.111830 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerName="glance-log" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.111853 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" containerName="glance-httpd" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.111865 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-log" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.111876 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" containerName="glance-httpd" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.112946 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.115684 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.116071 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.116149 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.116448 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bh9xs" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.121983 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.124827 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.130008 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.130318 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.132025 4809 scope.go:117] "RemoveContainer" containerID="174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.134507 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.155672 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.161242 4809 scope.go:117] "RemoveContainer" containerID="fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97" Nov 27 17:40:11 crc kubenswrapper[4809]: E1127 17:40:11.162271 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97\": container with ID starting with fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97 not found: ID does not exist" containerID="fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.162312 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97"} err="failed to get container status \"fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97\": rpc error: code = NotFound desc = could not find container \"fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97\": container with ID starting with fdb99317ba93ad811bc13d67d79c791e352bd1ba6ff8bd057628e24c7492ac97 not found: ID does not exist" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.162340 4809 scope.go:117] "RemoveContainer" containerID="174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045" Nov 27 17:40:11 crc kubenswrapper[4809]: E1127 17:40:11.162695 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045\": container with ID starting with 174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045 not found: ID does not exist" containerID="174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.162768 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045"} err="failed to get container status \"174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045\": rpc error: code = NotFound desc = could not find container \"174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045\": container with ID starting with 174a41dd063a242b9b5632a6e9c2e34ec7edeb4cebba937b4f5e7b3f235ff045 not found: ID does not exist" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187042 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187092 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187167 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-scripts\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187204 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/78ba6922-17ec-497d-b15d-e2ba250a698e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187222 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187253 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187284 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2339b0f8-362a-43b7-8679-8259e295d65e-logs\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187303 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2339b0f8-362a-43b7-8679-8259e295d65e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhpp7\" (UniqueName: \"kubernetes.io/projected/2339b0f8-362a-43b7-8679-8259e295d65e-kube-api-access-dhpp7\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187361 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187382 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gmqm\" (UniqueName: \"kubernetes.io/projected/78ba6922-17ec-497d-b15d-e2ba250a698e-kube-api-access-4gmqm\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187397 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187414 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78ba6922-17ec-497d-b15d-e2ba250a698e-logs\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.187429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-config-data\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78ba6922-17ec-497d-b15d-e2ba250a698e-logs\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289118 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-config-data\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289168 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289232 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289253 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289289 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-scripts\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289340 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/78ba6922-17ec-497d-b15d-e2ba250a698e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289369 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289411 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289530 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289590 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2339b0f8-362a-43b7-8679-8259e295d65e-logs\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289625 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2339b0f8-362a-43b7-8679-8259e295d65e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289646 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhpp7\" (UniqueName: \"kubernetes.io/projected/2339b0f8-362a-43b7-8679-8259e295d65e-kube-api-access-dhpp7\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289715 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gmqm\" (UniqueName: \"kubernetes.io/projected/78ba6922-17ec-497d-b15d-e2ba250a698e-kube-api-access-4gmqm\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289726 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78ba6922-17ec-497d-b15d-e2ba250a698e-logs\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289732 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.289964 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.290506 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2339b0f8-362a-43b7-8679-8259e295d65e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.290506 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2339b0f8-362a-43b7-8679-8259e295d65e-logs\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.293080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/78ba6922-17ec-497d-b15d-e2ba250a698e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.303300 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.304045 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.304620 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.305419 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.306913 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ba6922-17ec-497d-b15d-e2ba250a698e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.307179 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-scripts\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.308372 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.310630 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2339b0f8-362a-43b7-8679-8259e295d65e-config-data\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.311463 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhpp7\" (UniqueName: \"kubernetes.io/projected/2339b0f8-362a-43b7-8679-8259e295d65e-kube-api-access-dhpp7\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.312106 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gmqm\" (UniqueName: \"kubernetes.io/projected/78ba6922-17ec-497d-b15d-e2ba250a698e-kube-api-access-4gmqm\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.357302 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2339b0f8-362a-43b7-8679-8259e295d65e\") " pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.367238 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"78ba6922-17ec-497d-b15d-e2ba250a698e\") " pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.445013 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.461341 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.470337 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="596ec651-2088-4459-b0dd-224c67150b5d" path="/var/lib/kubelet/pods/596ec651-2088-4459-b0dd-224c67150b5d/volumes" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.471152 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f5d5190-5f50-4d4e-9739-250c85c0146b" path="/var/lib/kubelet/pods/7f5d5190-5f50-4d4e-9739-250c85c0146b/volumes" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.472362 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e345d6af-f198-42b0-9017-582c93ffeef6" path="/var/lib/kubelet/pods/e345d6af-f198-42b0-9017-582c93ffeef6/volumes" Nov 27 17:40:11 crc kubenswrapper[4809]: I1127 17:40:11.973606 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerStarted","Data":"000aa74db0465ad65ff866f7806acfcb93c1b0ce87b3bb15669b32859147d0bc"} Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.041024 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 17:40:12 crc kubenswrapper[4809]: W1127 17:40:12.052547 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2339b0f8_362a_43b7_8679_8259e295d65e.slice/crio-34dd79c2e7e0afcce0a75857bced14235c879aef5c06fbcf3c281057cbce832b WatchSource:0}: Error finding container 34dd79c2e7e0afcce0a75857bced14235c879aef5c06fbcf3c281057cbce832b: Status 404 returned error can't find the container with id 34dd79c2e7e0afcce0a75857bced14235c879aef5c06fbcf3c281057cbce832b Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.113609 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dnl76"] Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.114819 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.125476 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.125587 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hltnw" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.125788 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.133540 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dnl76"] Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.191767 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 17:40:12 crc kubenswrapper[4809]: W1127 17:40:12.209996 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78ba6922_17ec_497d_b15d_e2ba250a698e.slice/crio-47e347df53a37253118c1c6bdb62cb6161fad8540f992335c81d15e38ed4e931 WatchSource:0}: Error finding container 47e347df53a37253118c1c6bdb62cb6161fad8540f992335c81d15e38ed4e931: Status 404 returned error can't find the container with id 47e347df53a37253118c1c6bdb62cb6161fad8540f992335c81d15e38ed4e931 Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.212106 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-config-data\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.212206 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.212301 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7scfx\" (UniqueName: \"kubernetes.io/projected/b11296c7-060d-409c-852c-509694827f2f-kube-api-access-7scfx\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.212338 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-scripts\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.314060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7scfx\" (UniqueName: \"kubernetes.io/projected/b11296c7-060d-409c-852c-509694827f2f-kube-api-access-7scfx\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.314115 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-scripts\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.314177 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-config-data\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.314217 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.323139 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.323188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-config-data\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.325888 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-scripts\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.347780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7scfx\" (UniqueName: \"kubernetes.io/projected/b11296c7-060d-409c-852c-509694827f2f-kube-api-access-7scfx\") pod \"nova-cell0-conductor-db-sync-dnl76\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:12 crc kubenswrapper[4809]: I1127 17:40:12.491854 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:13 crc kubenswrapper[4809]: I1127 17:40:13.006692 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"78ba6922-17ec-497d-b15d-e2ba250a698e","Type":"ContainerStarted","Data":"456e874c9940d273c1a7fc9e9f535eca0bbcd13a2a353d6aa5b304edddeb99db"} Nov 27 17:40:13 crc kubenswrapper[4809]: I1127 17:40:13.007487 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"78ba6922-17ec-497d-b15d-e2ba250a698e","Type":"ContainerStarted","Data":"47e347df53a37253118c1c6bdb62cb6161fad8540f992335c81d15e38ed4e931"} Nov 27 17:40:13 crc kubenswrapper[4809]: I1127 17:40:13.010553 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2339b0f8-362a-43b7-8679-8259e295d65e","Type":"ContainerStarted","Data":"ca34d2c5bf45152048fe12f32ce4b2da05f31424774fe4f6ae23ff979d575625"} Nov 27 17:40:13 crc kubenswrapper[4809]: I1127 17:40:13.010587 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2339b0f8-362a-43b7-8679-8259e295d65e","Type":"ContainerStarted","Data":"34dd79c2e7e0afcce0a75857bced14235c879aef5c06fbcf3c281057cbce832b"} Nov 27 17:40:13 crc kubenswrapper[4809]: I1127 17:40:13.013015 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerStarted","Data":"dd2559b6ac34d9807ca53abfa93d3d7b25b2ed59983629104549a9a253d6c78c"} Nov 27 17:40:13 crc kubenswrapper[4809]: I1127 17:40:13.013050 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerStarted","Data":"d58debb2d0143b5b0ee5384d38c5b611c4761dcff6e8c468a3602ba5575b2c64"} Nov 27 17:40:13 crc kubenswrapper[4809]: I1127 17:40:13.138837 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dnl76"] Nov 27 17:40:14 crc kubenswrapper[4809]: I1127 17:40:14.038285 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dnl76" event={"ID":"b11296c7-060d-409c-852c-509694827f2f","Type":"ContainerStarted","Data":"65f4d7cd5c047d957b12fad0c3a2f97a2ed01d8336a9d7c789974397b2cd5fc5"} Nov 27 17:40:14 crc kubenswrapper[4809]: I1127 17:40:14.046503 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerStarted","Data":"43036a15429b6497abb2b720fc10ad9a5ff93f882a24b3ae885a1aacc2ac0b2c"} Nov 27 17:40:14 crc kubenswrapper[4809]: I1127 17:40:14.055721 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"78ba6922-17ec-497d-b15d-e2ba250a698e","Type":"ContainerStarted","Data":"d89b77b1623dd6fafa509dacd6bfc922039d33eeb7eeee45d4d8d99c9dead792"} Nov 27 17:40:14 crc kubenswrapper[4809]: I1127 17:40:14.058115 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2339b0f8-362a-43b7-8679-8259e295d65e","Type":"ContainerStarted","Data":"9fc88cbcd9147ee6f80aa30f8222e7f0d6d8219d3d2f4fa253bb122c4937f160"} Nov 27 17:40:14 crc kubenswrapper[4809]: I1127 17:40:14.086568 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.086544147 podStartE2EDuration="3.086544147s" podCreationTimestamp="2025-11-27 17:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:14.073721029 +0000 UTC m=+1849.346178381" watchObservedRunningTime="2025-11-27 17:40:14.086544147 +0000 UTC m=+1849.359001509" Nov 27 17:40:14 crc kubenswrapper[4809]: I1127 17:40:14.101406 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.101391022 podStartE2EDuration="3.101391022s" podCreationTimestamp="2025-11-27 17:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:14.094223927 +0000 UTC m=+1849.366681279" watchObservedRunningTime="2025-11-27 17:40:14.101391022 +0000 UTC m=+1849.373848374" Nov 27 17:40:14 crc kubenswrapper[4809]: I1127 17:40:14.458631 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:40:14 crc kubenswrapper[4809]: E1127 17:40:14.459279 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:40:15 crc kubenswrapper[4809]: I1127 17:40:15.073331 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="ceilometer-central-agent" containerID="cri-o://d58debb2d0143b5b0ee5384d38c5b611c4761dcff6e8c468a3602ba5575b2c64" gracePeriod=30 Nov 27 17:40:15 crc kubenswrapper[4809]: I1127 17:40:15.073726 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerStarted","Data":"82d594d2801ba34b53a05e29b81e5bbf12b817911e26836f3f88c3965d9d887b"} Nov 27 17:40:15 crc kubenswrapper[4809]: I1127 17:40:15.073827 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 17:40:15 crc kubenswrapper[4809]: I1127 17:40:15.073992 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="sg-core" containerID="cri-o://43036a15429b6497abb2b720fc10ad9a5ff93f882a24b3ae885a1aacc2ac0b2c" gracePeriod=30 Nov 27 17:40:15 crc kubenswrapper[4809]: I1127 17:40:15.074053 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="proxy-httpd" containerID="cri-o://82d594d2801ba34b53a05e29b81e5bbf12b817911e26836f3f88c3965d9d887b" gracePeriod=30 Nov 27 17:40:15 crc kubenswrapper[4809]: I1127 17:40:15.074142 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="ceilometer-notification-agent" containerID="cri-o://dd2559b6ac34d9807ca53abfa93d3d7b25b2ed59983629104549a9a253d6c78c" gracePeriod=30 Nov 27 17:40:15 crc kubenswrapper[4809]: I1127 17:40:15.102003 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.43213314 podStartE2EDuration="6.101987136s" podCreationTimestamp="2025-11-27 17:40:09 +0000 UTC" firstStartedPulling="2025-11-27 17:40:11.048312023 +0000 UTC m=+1846.320769375" lastFinishedPulling="2025-11-27 17:40:14.718166019 +0000 UTC m=+1849.990623371" observedRunningTime="2025-11-27 17:40:15.091753377 +0000 UTC m=+1850.364210729" watchObservedRunningTime="2025-11-27 17:40:15.101987136 +0000 UTC m=+1850.374444488" Nov 27 17:40:16 crc kubenswrapper[4809]: I1127 17:40:16.088262 4809 generic.go:334] "Generic (PLEG): container finished" podID="3878cd07-3090-409e-add5-6d020076b63b" containerID="82d594d2801ba34b53a05e29b81e5bbf12b817911e26836f3f88c3965d9d887b" exitCode=0 Nov 27 17:40:16 crc kubenswrapper[4809]: I1127 17:40:16.088521 4809 generic.go:334] "Generic (PLEG): container finished" podID="3878cd07-3090-409e-add5-6d020076b63b" containerID="43036a15429b6497abb2b720fc10ad9a5ff93f882a24b3ae885a1aacc2ac0b2c" exitCode=2 Nov 27 17:40:16 crc kubenswrapper[4809]: I1127 17:40:16.088531 4809 generic.go:334] "Generic (PLEG): container finished" podID="3878cd07-3090-409e-add5-6d020076b63b" containerID="dd2559b6ac34d9807ca53abfa93d3d7b25b2ed59983629104549a9a253d6c78c" exitCode=0 Nov 27 17:40:16 crc kubenswrapper[4809]: I1127 17:40:16.088355 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerDied","Data":"82d594d2801ba34b53a05e29b81e5bbf12b817911e26836f3f88c3965d9d887b"} Nov 27 17:40:16 crc kubenswrapper[4809]: I1127 17:40:16.088565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerDied","Data":"43036a15429b6497abb2b720fc10ad9a5ff93f882a24b3ae885a1aacc2ac0b2c"} Nov 27 17:40:16 crc kubenswrapper[4809]: I1127 17:40:16.088578 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerDied","Data":"dd2559b6ac34d9807ca53abfa93d3d7b25b2ed59983629104549a9a253d6c78c"} Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.130953 4809 generic.go:334] "Generic (PLEG): container finished" podID="3878cd07-3090-409e-add5-6d020076b63b" containerID="d58debb2d0143b5b0ee5384d38c5b611c4761dcff6e8c468a3602ba5575b2c64" exitCode=0 Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.131024 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerDied","Data":"d58debb2d0143b5b0ee5384d38c5b611c4761dcff6e8c468a3602ba5575b2c64"} Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.391110 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.511444 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-combined-ca-bundle\") pod \"3878cd07-3090-409e-add5-6d020076b63b\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.511573 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-sg-core-conf-yaml\") pod \"3878cd07-3090-409e-add5-6d020076b63b\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.511660 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-run-httpd\") pod \"3878cd07-3090-409e-add5-6d020076b63b\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.511803 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwnpx\" (UniqueName: \"kubernetes.io/projected/3878cd07-3090-409e-add5-6d020076b63b-kube-api-access-mwnpx\") pod \"3878cd07-3090-409e-add5-6d020076b63b\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.511841 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-config-data\") pod \"3878cd07-3090-409e-add5-6d020076b63b\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.511878 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-log-httpd\") pod \"3878cd07-3090-409e-add5-6d020076b63b\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.511965 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-scripts\") pod \"3878cd07-3090-409e-add5-6d020076b63b\" (UID: \"3878cd07-3090-409e-add5-6d020076b63b\") " Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.512185 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3878cd07-3090-409e-add5-6d020076b63b" (UID: "3878cd07-3090-409e-add5-6d020076b63b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.512490 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3878cd07-3090-409e-add5-6d020076b63b" (UID: "3878cd07-3090-409e-add5-6d020076b63b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.512820 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.512841 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3878cd07-3090-409e-add5-6d020076b63b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.516615 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3878cd07-3090-409e-add5-6d020076b63b-kube-api-access-mwnpx" (OuterVolumeSpecName: "kube-api-access-mwnpx") pod "3878cd07-3090-409e-add5-6d020076b63b" (UID: "3878cd07-3090-409e-add5-6d020076b63b"). InnerVolumeSpecName "kube-api-access-mwnpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.516612 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-scripts" (OuterVolumeSpecName: "scripts") pod "3878cd07-3090-409e-add5-6d020076b63b" (UID: "3878cd07-3090-409e-add5-6d020076b63b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.544206 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3878cd07-3090-409e-add5-6d020076b63b" (UID: "3878cd07-3090-409e-add5-6d020076b63b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.585959 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3878cd07-3090-409e-add5-6d020076b63b" (UID: "3878cd07-3090-409e-add5-6d020076b63b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.609802 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-config-data" (OuterVolumeSpecName: "config-data") pod "3878cd07-3090-409e-add5-6d020076b63b" (UID: "3878cd07-3090-409e-add5-6d020076b63b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.615106 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.615135 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.615145 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.615156 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3878cd07-3090-409e-add5-6d020076b63b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:20 crc kubenswrapper[4809]: I1127 17:40:20.615166 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwnpx\" (UniqueName: \"kubernetes.io/projected/3878cd07-3090-409e-add5-6d020076b63b-kube-api-access-mwnpx\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.145237 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dnl76" event={"ID":"b11296c7-060d-409c-852c-509694827f2f","Type":"ContainerStarted","Data":"6f4fe92641ae06bfe13cdd438b6db28a6cbc7ea22c125760d0fb468229e1a742"} Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.155448 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3878cd07-3090-409e-add5-6d020076b63b","Type":"ContainerDied","Data":"000aa74db0465ad65ff866f7806acfcb93c1b0ce87b3bb15669b32859147d0bc"} Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.155828 4809 scope.go:117] "RemoveContainer" containerID="82d594d2801ba34b53a05e29b81e5bbf12b817911e26836f3f88c3965d9d887b" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.155995 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.171168 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-dnl76" podStartSLOduration=2.071808137 podStartE2EDuration="9.171145517s" podCreationTimestamp="2025-11-27 17:40:12 +0000 UTC" firstStartedPulling="2025-11-27 17:40:13.149527064 +0000 UTC m=+1848.421984416" lastFinishedPulling="2025-11-27 17:40:20.248864444 +0000 UTC m=+1855.521321796" observedRunningTime="2025-11-27 17:40:21.168384861 +0000 UTC m=+1856.440842223" watchObservedRunningTime="2025-11-27 17:40:21.171145517 +0000 UTC m=+1856.443602879" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.198351 4809 scope.go:117] "RemoveContainer" containerID="43036a15429b6497abb2b720fc10ad9a5ff93f882a24b3ae885a1aacc2ac0b2c" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.203942 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.214514 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.226199 4809 scope.go:117] "RemoveContainer" containerID="dd2559b6ac34d9807ca53abfa93d3d7b25b2ed59983629104549a9a253d6c78c" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.226558 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:21 crc kubenswrapper[4809]: E1127 17:40:21.227031 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="proxy-httpd" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.227044 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="proxy-httpd" Nov 27 17:40:21 crc kubenswrapper[4809]: E1127 17:40:21.227079 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="ceilometer-notification-agent" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.227086 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="ceilometer-notification-agent" Nov 27 17:40:21 crc kubenswrapper[4809]: E1127 17:40:21.227104 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="sg-core" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.227111 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="sg-core" Nov 27 17:40:21 crc kubenswrapper[4809]: E1127 17:40:21.227128 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="ceilometer-central-agent" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.227135 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="ceilometer-central-agent" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.227324 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="sg-core" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.227334 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="ceilometer-central-agent" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.227347 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="ceilometer-notification-agent" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.227356 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3878cd07-3090-409e-add5-6d020076b63b" containerName="proxy-httpd" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.229203 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.232150 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.232367 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.253796 4809 scope.go:117] "RemoveContainer" containerID="d58debb2d0143b5b0ee5384d38c5b611c4761dcff6e8c468a3602ba5575b2c64" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.263193 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.329123 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-run-httpd\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.329201 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-scripts\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.329220 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-log-httpd\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.329622 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs8m2\" (UniqueName: \"kubernetes.io/projected/ef77ec6b-da7e-4c88-8861-25bc3d925a08-kube-api-access-rs8m2\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.329715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-config-data\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.329811 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.329894 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.431923 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-scripts\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.431961 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-log-httpd\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.432061 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs8m2\" (UniqueName: \"kubernetes.io/projected/ef77ec6b-da7e-4c88-8861-25bc3d925a08-kube-api-access-rs8m2\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.432087 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-config-data\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.432119 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.432149 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.432362 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-run-httpd\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.432833 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-log-httpd\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.433617 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-run-httpd\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.437021 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.438064 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-config-data\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.440973 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.442815 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-scripts\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.447442 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.447904 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.458548 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs8m2\" (UniqueName: \"kubernetes.io/projected/ef77ec6b-da7e-4c88-8861-25bc3d925a08-kube-api-access-rs8m2\") pod \"ceilometer-0\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " pod="openstack/ceilometer-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.470891 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3878cd07-3090-409e-add5-6d020076b63b" path="/var/lib/kubelet/pods/3878cd07-3090-409e-add5-6d020076b63b/volumes" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.472099 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.472142 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.488350 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.502540 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.502700 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.512837 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:21 crc kubenswrapper[4809]: I1127 17:40:21.557508 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:22 crc kubenswrapper[4809]: I1127 17:40:22.026314 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:22 crc kubenswrapper[4809]: I1127 17:40:22.177515 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerStarted","Data":"a4c93108a51da792bc3d7b0c8501a0fdbb1e656719897e54386214d249ad88ad"} Nov 27 17:40:22 crc kubenswrapper[4809]: I1127 17:40:22.179731 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 27 17:40:22 crc kubenswrapper[4809]: I1127 17:40:22.179784 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:22 crc kubenswrapper[4809]: I1127 17:40:22.179795 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:22 crc kubenswrapper[4809]: I1127 17:40:22.179804 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 27 17:40:23 crc kubenswrapper[4809]: I1127 17:40:23.453964 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerStarted","Data":"ee726de285c697cea3d6f0da36893db3cad3a34ac0297ae833372608b4051769"} Nov 27 17:40:23 crc kubenswrapper[4809]: I1127 17:40:23.950883 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:24 crc kubenswrapper[4809]: I1127 17:40:24.428937 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 27 17:40:24 crc kubenswrapper[4809]: I1127 17:40:24.435548 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:24 crc kubenswrapper[4809]: I1127 17:40:24.439672 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 27 17:40:24 crc kubenswrapper[4809]: I1127 17:40:24.478372 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 17:40:24 crc kubenswrapper[4809]: I1127 17:40:24.478407 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerStarted","Data":"ca989654d52d9f27f8960140c0f9b164217d4965b07245084908bbea6d1209c9"} Nov 27 17:40:24 crc kubenswrapper[4809]: I1127 17:40:24.609867 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 27 17:40:25 crc kubenswrapper[4809]: I1127 17:40:25.503313 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerStarted","Data":"48f127e299f21007b78ba91b7a296eb2e2b8abc2bbb1edf51f71a2ee8988ddbe"} Nov 27 17:40:26 crc kubenswrapper[4809]: I1127 17:40:26.515628 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerStarted","Data":"8d2c19dd91edfc4eb517c27e678cb29882d726a7bd76b7b3c9a80ab7ec08e852"} Nov 27 17:40:26 crc kubenswrapper[4809]: I1127 17:40:26.516133 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="ceilometer-central-agent" containerID="cri-o://ee726de285c697cea3d6f0da36893db3cad3a34ac0297ae833372608b4051769" gracePeriod=30 Nov 27 17:40:26 crc kubenswrapper[4809]: I1127 17:40:26.516428 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 17:40:26 crc kubenswrapper[4809]: I1127 17:40:26.516765 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="proxy-httpd" containerID="cri-o://8d2c19dd91edfc4eb517c27e678cb29882d726a7bd76b7b3c9a80ab7ec08e852" gracePeriod=30 Nov 27 17:40:26 crc kubenswrapper[4809]: I1127 17:40:26.516824 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="sg-core" containerID="cri-o://48f127e299f21007b78ba91b7a296eb2e2b8abc2bbb1edf51f71a2ee8988ddbe" gracePeriod=30 Nov 27 17:40:26 crc kubenswrapper[4809]: I1127 17:40:26.516871 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="ceilometer-notification-agent" containerID="cri-o://ca989654d52d9f27f8960140c0f9b164217d4965b07245084908bbea6d1209c9" gracePeriod=30 Nov 27 17:40:26 crc kubenswrapper[4809]: I1127 17:40:26.556692 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.429252921 podStartE2EDuration="5.556672311s" podCreationTimestamp="2025-11-27 17:40:21 +0000 UTC" firstStartedPulling="2025-11-27 17:40:22.031775792 +0000 UTC m=+1857.304233144" lastFinishedPulling="2025-11-27 17:40:26.159195172 +0000 UTC m=+1861.431652534" observedRunningTime="2025-11-27 17:40:26.551019317 +0000 UTC m=+1861.823476669" watchObservedRunningTime="2025-11-27 17:40:26.556672311 +0000 UTC m=+1861.829129663" Nov 27 17:40:27 crc kubenswrapper[4809]: I1127 17:40:27.458687 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:40:27 crc kubenswrapper[4809]: E1127 17:40:27.459003 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:40:27 crc kubenswrapper[4809]: I1127 17:40:27.526329 4809 generic.go:334] "Generic (PLEG): container finished" podID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerID="8d2c19dd91edfc4eb517c27e678cb29882d726a7bd76b7b3c9a80ab7ec08e852" exitCode=0 Nov 27 17:40:27 crc kubenswrapper[4809]: I1127 17:40:27.526365 4809 generic.go:334] "Generic (PLEG): container finished" podID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerID="48f127e299f21007b78ba91b7a296eb2e2b8abc2bbb1edf51f71a2ee8988ddbe" exitCode=2 Nov 27 17:40:27 crc kubenswrapper[4809]: I1127 17:40:27.526372 4809 generic.go:334] "Generic (PLEG): container finished" podID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerID="ca989654d52d9f27f8960140c0f9b164217d4965b07245084908bbea6d1209c9" exitCode=0 Nov 27 17:40:27 crc kubenswrapper[4809]: I1127 17:40:27.526392 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerDied","Data":"8d2c19dd91edfc4eb517c27e678cb29882d726a7bd76b7b3c9a80ab7ec08e852"} Nov 27 17:40:27 crc kubenswrapper[4809]: I1127 17:40:27.526419 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerDied","Data":"48f127e299f21007b78ba91b7a296eb2e2b8abc2bbb1edf51f71a2ee8988ddbe"} Nov 27 17:40:27 crc kubenswrapper[4809]: I1127 17:40:27.526429 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerDied","Data":"ca989654d52d9f27f8960140c0f9b164217d4965b07245084908bbea6d1209c9"} Nov 27 17:40:31 crc kubenswrapper[4809]: I1127 17:40:31.563653 4809 generic.go:334] "Generic (PLEG): container finished" podID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerID="ee726de285c697cea3d6f0da36893db3cad3a34ac0297ae833372608b4051769" exitCode=0 Nov 27 17:40:31 crc kubenswrapper[4809]: I1127 17:40:31.563685 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerDied","Data":"ee726de285c697cea3d6f0da36893db3cad3a34ac0297ae833372608b4051769"} Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.024998 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.171179 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-config-data\") pod \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.171270 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs8m2\" (UniqueName: \"kubernetes.io/projected/ef77ec6b-da7e-4c88-8861-25bc3d925a08-kube-api-access-rs8m2\") pod \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.171326 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-combined-ca-bundle\") pod \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.171374 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-scripts\") pod \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.171419 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-sg-core-conf-yaml\") pod \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.171463 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-log-httpd\") pod \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.171492 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-run-httpd\") pod \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\" (UID: \"ef77ec6b-da7e-4c88-8861-25bc3d925a08\") " Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.172162 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ef77ec6b-da7e-4c88-8861-25bc3d925a08" (UID: "ef77ec6b-da7e-4c88-8861-25bc3d925a08"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.172481 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ef77ec6b-da7e-4c88-8861-25bc3d925a08" (UID: "ef77ec6b-da7e-4c88-8861-25bc3d925a08"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.178302 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef77ec6b-da7e-4c88-8861-25bc3d925a08-kube-api-access-rs8m2" (OuterVolumeSpecName: "kube-api-access-rs8m2") pod "ef77ec6b-da7e-4c88-8861-25bc3d925a08" (UID: "ef77ec6b-da7e-4c88-8861-25bc3d925a08"). InnerVolumeSpecName "kube-api-access-rs8m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.186893 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-scripts" (OuterVolumeSpecName: "scripts") pod "ef77ec6b-da7e-4c88-8861-25bc3d925a08" (UID: "ef77ec6b-da7e-4c88-8861-25bc3d925a08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.205259 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ef77ec6b-da7e-4c88-8861-25bc3d925a08" (UID: "ef77ec6b-da7e-4c88-8861-25bc3d925a08"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.246860 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef77ec6b-da7e-4c88-8861-25bc3d925a08" (UID: "ef77ec6b-da7e-4c88-8861-25bc3d925a08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.266923 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-config-data" (OuterVolumeSpecName: "config-data") pod "ef77ec6b-da7e-4c88-8861-25bc3d925a08" (UID: "ef77ec6b-da7e-4c88-8861-25bc3d925a08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.273964 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.274007 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef77ec6b-da7e-4c88-8861-25bc3d925a08-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.274019 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.274063 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs8m2\" (UniqueName: \"kubernetes.io/projected/ef77ec6b-da7e-4c88-8861-25bc3d925a08-kube-api-access-rs8m2\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.274077 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.274104 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.274114 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef77ec6b-da7e-4c88-8861-25bc3d925a08-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.593129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef77ec6b-da7e-4c88-8861-25bc3d925a08","Type":"ContainerDied","Data":"a4c93108a51da792bc3d7b0c8501a0fdbb1e656719897e54386214d249ad88ad"} Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.593192 4809 scope.go:117] "RemoveContainer" containerID="8d2c19dd91edfc4eb517c27e678cb29882d726a7bd76b7b3c9a80ab7ec08e852" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.593202 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.637437 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.639071 4809 scope.go:117] "RemoveContainer" containerID="48f127e299f21007b78ba91b7a296eb2e2b8abc2bbb1edf51f71a2ee8988ddbe" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.649349 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.659505 4809 scope.go:117] "RemoveContainer" containerID="ca989654d52d9f27f8960140c0f9b164217d4965b07245084908bbea6d1209c9" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.660165 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:34 crc kubenswrapper[4809]: E1127 17:40:34.660693 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="sg-core" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.660720 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="sg-core" Nov 27 17:40:34 crc kubenswrapper[4809]: E1127 17:40:34.660757 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="ceilometer-notification-agent" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.660766 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="ceilometer-notification-agent" Nov 27 17:40:34 crc kubenswrapper[4809]: E1127 17:40:34.660791 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="proxy-httpd" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.660800 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="proxy-httpd" Nov 27 17:40:34 crc kubenswrapper[4809]: E1127 17:40:34.660847 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="ceilometer-central-agent" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.660858 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="ceilometer-central-agent" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.661085 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="ceilometer-central-agent" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.661120 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="sg-core" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.661151 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="ceilometer-notification-agent" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.661162 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" containerName="proxy-httpd" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.663331 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.667953 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.668504 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.674272 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.699786 4809 scope.go:117] "RemoveContainer" containerID="ee726de285c697cea3d6f0da36893db3cad3a34ac0297ae833372608b4051769" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.783429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-config-data\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.783510 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.783592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.783670 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-log-httpd\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.783808 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-scripts\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.783859 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-run-httpd\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.784131 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6mht\" (UniqueName: \"kubernetes.io/projected/f802a83a-cdca-4a71-911c-63dc4c8f14f6-kube-api-access-r6mht\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.885852 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.885951 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-log-httpd\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.885993 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-scripts\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.886015 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-run-httpd\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.886131 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6mht\" (UniqueName: \"kubernetes.io/projected/f802a83a-cdca-4a71-911c-63dc4c8f14f6-kube-api-access-r6mht\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.886172 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-config-data\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.886215 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.886551 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-log-httpd\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.886580 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-run-httpd\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.891200 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.891239 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-scripts\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.891280 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.892846 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-config-data\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.908144 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6mht\" (UniqueName: \"kubernetes.io/projected/f802a83a-cdca-4a71-911c-63dc4c8f14f6-kube-api-access-r6mht\") pod \"ceilometer-0\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " pod="openstack/ceilometer-0" Nov 27 17:40:34 crc kubenswrapper[4809]: I1127 17:40:34.990153 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:40:35 crc kubenswrapper[4809]: I1127 17:40:35.426369 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:40:35 crc kubenswrapper[4809]: I1127 17:40:35.479218 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef77ec6b-da7e-4c88-8861-25bc3d925a08" path="/var/lib/kubelet/pods/ef77ec6b-da7e-4c88-8861-25bc3d925a08/volumes" Nov 27 17:40:35 crc kubenswrapper[4809]: I1127 17:40:35.603941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerStarted","Data":"f7c7348e8ada464abee6949902a178e7a365dcbac8e78c4068f6ac3767000b8f"} Nov 27 17:40:35 crc kubenswrapper[4809]: I1127 17:40:35.605370 4809 generic.go:334] "Generic (PLEG): container finished" podID="b11296c7-060d-409c-852c-509694827f2f" containerID="6f4fe92641ae06bfe13cdd438b6db28a6cbc7ea22c125760d0fb468229e1a742" exitCode=0 Nov 27 17:40:35 crc kubenswrapper[4809]: I1127 17:40:35.605436 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dnl76" event={"ID":"b11296c7-060d-409c-852c-509694827f2f","Type":"ContainerDied","Data":"6f4fe92641ae06bfe13cdd438b6db28a6cbc7ea22c125760d0fb468229e1a742"} Nov 27 17:40:36 crc kubenswrapper[4809]: I1127 17:40:36.638783 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerStarted","Data":"c52a6040f3804de9369eccd5891abcf98bff5e7d1b87d9f9a1ceb591946cad99"} Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.042990 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.137673 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-config-data\") pod \"b11296c7-060d-409c-852c-509694827f2f\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.137798 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-scripts\") pod \"b11296c7-060d-409c-852c-509694827f2f\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.137950 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-combined-ca-bundle\") pod \"b11296c7-060d-409c-852c-509694827f2f\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.138060 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7scfx\" (UniqueName: \"kubernetes.io/projected/b11296c7-060d-409c-852c-509694827f2f-kube-api-access-7scfx\") pod \"b11296c7-060d-409c-852c-509694827f2f\" (UID: \"b11296c7-060d-409c-852c-509694827f2f\") " Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.146657 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11296c7-060d-409c-852c-509694827f2f-kube-api-access-7scfx" (OuterVolumeSpecName: "kube-api-access-7scfx") pod "b11296c7-060d-409c-852c-509694827f2f" (UID: "b11296c7-060d-409c-852c-509694827f2f"). InnerVolumeSpecName "kube-api-access-7scfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.161328 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-scripts" (OuterVolumeSpecName: "scripts") pod "b11296c7-060d-409c-852c-509694827f2f" (UID: "b11296c7-060d-409c-852c-509694827f2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.174548 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b11296c7-060d-409c-852c-509694827f2f" (UID: "b11296c7-060d-409c-852c-509694827f2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.194264 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-config-data" (OuterVolumeSpecName: "config-data") pod "b11296c7-060d-409c-852c-509694827f2f" (UID: "b11296c7-060d-409c-852c-509694827f2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.240835 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.240874 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7scfx\" (UniqueName: \"kubernetes.io/projected/b11296c7-060d-409c-852c-509694827f2f-kube-api-access-7scfx\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.240886 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.240895 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11296c7-060d-409c-852c-509694827f2f-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.653491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerStarted","Data":"bd8eec56078ff1a5910c47c787dadb501a8f077b8412160d3774e70541b98b24"} Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.653816 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerStarted","Data":"162fdb74efc4fa762b0e1378da20e568bcd84b3daeea002661753778e9f08150"} Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.659370 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dnl76" event={"ID":"b11296c7-060d-409c-852c-509694827f2f","Type":"ContainerDied","Data":"65f4d7cd5c047d957b12fad0c3a2f97a2ed01d8336a9d7c789974397b2cd5fc5"} Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.659567 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65f4d7cd5c047d957b12fad0c3a2f97a2ed01d8336a9d7c789974397b2cd5fc5" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.659413 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dnl76" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.729924 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 27 17:40:37 crc kubenswrapper[4809]: E1127 17:40:37.730341 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11296c7-060d-409c-852c-509694827f2f" containerName="nova-cell0-conductor-db-sync" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.730357 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11296c7-060d-409c-852c-509694827f2f" containerName="nova-cell0-conductor-db-sync" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.730553 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11296c7-060d-409c-852c-509694827f2f" containerName="nova-cell0-conductor-db-sync" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.731197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.733368 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hltnw" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.739209 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.743830 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.853614 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75rcc\" (UniqueName: \"kubernetes.io/projected/4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b-kube-api-access-75rcc\") pod \"nova-cell0-conductor-0\" (UID: \"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.853983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.854171 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.956548 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.956636 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.956753 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75rcc\" (UniqueName: \"kubernetes.io/projected/4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b-kube-api-access-75rcc\") pod \"nova-cell0-conductor-0\" (UID: \"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.960381 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.960486 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:37 crc kubenswrapper[4809]: I1127 17:40:37.978874 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75rcc\" (UniqueName: \"kubernetes.io/projected/4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b-kube-api-access-75rcc\") pod \"nova-cell0-conductor-0\" (UID: \"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b\") " pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:38 crc kubenswrapper[4809]: I1127 17:40:38.046765 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:38 crc kubenswrapper[4809]: I1127 17:40:38.498197 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 27 17:40:38 crc kubenswrapper[4809]: I1127 17:40:38.670507 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b","Type":"ContainerStarted","Data":"399dd4b5a4421de62089789fa0399d4559a945a6abe2c7a8c3f6a4de6a928186"} Nov 27 17:40:38 crc kubenswrapper[4809]: I1127 17:40:38.671014 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:38 crc kubenswrapper[4809]: I1127 17:40:38.690938 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.690916571 podStartE2EDuration="1.690916571s" podCreationTimestamp="2025-11-27 17:40:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:38.686367107 +0000 UTC m=+1873.958824459" watchObservedRunningTime="2025-11-27 17:40:38.690916571 +0000 UTC m=+1873.963373923" Nov 27 17:40:39 crc kubenswrapper[4809]: I1127 17:40:39.682358 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerStarted","Data":"4642b43a93b98e574260a834a13c519c1eda1ddb903a08eccc2d908cada91d75"} Nov 27 17:40:39 crc kubenswrapper[4809]: I1127 17:40:39.684046 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 17:40:39 crc kubenswrapper[4809]: I1127 17:40:39.686152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b","Type":"ContainerStarted","Data":"940e68d626de4aa31fff0f418c5b0affe9bc7f91ebd51d74ba2e444c4e6d04a1"} Nov 27 17:40:39 crc kubenswrapper[4809]: I1127 17:40:39.716103 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.409461363 podStartE2EDuration="5.716077593s" podCreationTimestamp="2025-11-27 17:40:34 +0000 UTC" firstStartedPulling="2025-11-27 17:40:35.434001684 +0000 UTC m=+1870.706459036" lastFinishedPulling="2025-11-27 17:40:38.740617914 +0000 UTC m=+1874.013075266" observedRunningTime="2025-11-27 17:40:39.707067458 +0000 UTC m=+1874.979524820" watchObservedRunningTime="2025-11-27 17:40:39.716077593 +0000 UTC m=+1874.988534965" Nov 27 17:40:42 crc kubenswrapper[4809]: I1127 17:40:42.458058 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:40:42 crc kubenswrapper[4809]: E1127 17:40:42.458800 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.080491 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.553948 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-wpb8b"] Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.555558 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.557709 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.558478 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.565510 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-wpb8b"] Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.686298 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.686619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-config-data\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.686665 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzzrc\" (UniqueName: \"kubernetes.io/projected/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-kube-api-access-vzzrc\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.686715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-scripts\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.766371 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.768126 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.774874 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.788646 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-config-data\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.788693 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzzrc\" (UniqueName: \"kubernetes.io/projected/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-kube-api-access-vzzrc\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.788729 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-scripts\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.788780 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.793882 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.799420 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.800297 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-config-data\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.808802 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-scripts\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.814712 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzzrc\" (UniqueName: \"kubernetes.io/projected/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-kube-api-access-vzzrc\") pod \"nova-cell0-cell-mapping-wpb8b\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.883585 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.890423 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-config-data\") pod \"nova-scheduler-0\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " pod="openstack/nova-scheduler-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.890473 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5wfz\" (UniqueName: \"kubernetes.io/projected/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-kube-api-access-x5wfz\") pod \"nova-scheduler-0\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " pod="openstack/nova-scheduler-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.890509 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " pod="openstack/nova-scheduler-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.909395 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.916305 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.924251 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.944447 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.997619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-config-data\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.997720 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvw2x\" (UniqueName: \"kubernetes.io/projected/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-kube-api-access-zvw2x\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.997791 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.997827 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-config-data\") pod \"nova-scheduler-0\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " pod="openstack/nova-scheduler-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.997844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5wfz\" (UniqueName: \"kubernetes.io/projected/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-kube-api-access-x5wfz\") pod \"nova-scheduler-0\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " pod="openstack/nova-scheduler-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.997866 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " pod="openstack/nova-scheduler-0" Nov 27 17:40:43 crc kubenswrapper[4809]: I1127 17:40:43.997910 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-logs\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.007478 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-config-data\") pod \"nova-scheduler-0\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " pod="openstack/nova-scheduler-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.011820 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " pod="openstack/nova-scheduler-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.053617 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5wfz\" (UniqueName: \"kubernetes.io/projected/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-kube-api-access-x5wfz\") pod \"nova-scheduler-0\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " pod="openstack/nova-scheduler-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.089866 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.094384 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.109255 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-config-data\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.109342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvw2x\" (UniqueName: \"kubernetes.io/projected/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-kube-api-access-zvw2x\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.109407 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.109490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-logs\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.115394 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-logs\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.125376 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-config-data\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.128620 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.135085 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.135121 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.135976 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.136946 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.137706 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvw2x\" (UniqueName: \"kubernetes.io/projected/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-kube-api-access-zvw2x\") pod \"nova-metadata-0\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.139778 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xtlg4"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.140242 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.141898 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.145607 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.173073 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.201555 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xtlg4"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.211543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.211630 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-config-data\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.211681 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.211723 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlnjx\" (UniqueName: \"kubernetes.io/projected/d3c6e105-a2d2-4016-b84d-4444f89fb22f-kube-api-access-vlnjx\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.211843 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3c6e105-a2d2-4016-b84d-4444f89fb22f-logs\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.211866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n7rk\" (UniqueName: \"kubernetes.io/projected/0ffdcceb-d220-4a44-88f9-94f7caf3694f-kube-api-access-6n7rk\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.211894 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314107 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314163 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlnjx\" (UniqueName: \"kubernetes.io/projected/d3c6e105-a2d2-4016-b84d-4444f89fb22f-kube-api-access-vlnjx\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314222 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3c6e105-a2d2-4016-b84d-4444f89fb22f-logs\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314242 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n7rk\" (UniqueName: \"kubernetes.io/projected/0ffdcceb-d220-4a44-88f9-94f7caf3694f-kube-api-access-6n7rk\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314268 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314289 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314336 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314411 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-config\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314451 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27cjl\" (UniqueName: \"kubernetes.io/projected/16fbb102-acdd-49fd-a941-9e7ff32b71cc-kube-api-access-27cjl\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.314485 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-config-data\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.315265 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3c6e105-a2d2-4016-b84d-4444f89fb22f-logs\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.319863 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.323288 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.325651 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.326405 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-config-data\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.367300 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlnjx\" (UniqueName: \"kubernetes.io/projected/d3c6e105-a2d2-4016-b84d-4444f89fb22f-kube-api-access-vlnjx\") pod \"nova-api-0\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.389337 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n7rk\" (UniqueName: \"kubernetes.io/projected/0ffdcceb-d220-4a44-88f9-94f7caf3694f-kube-api-access-6n7rk\") pod \"nova-cell1-novncproxy-0\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.416526 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.415733 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.416638 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-config\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.416657 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27cjl\" (UniqueName: \"kubernetes.io/projected/16fbb102-acdd-49fd-a941-9e7ff32b71cc-kube-api-access-27cjl\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.417099 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.417806 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.417133 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.417880 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.418012 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.418494 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.418558 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-config\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.427273 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.432732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27cjl\" (UniqueName: \"kubernetes.io/projected/16fbb102-acdd-49fd-a941-9e7ff32b71cc-kube-api-access-27cjl\") pod \"dnsmasq-dns-845d6d6f59-xtlg4\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.471372 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.510801 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.515177 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.683004 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-wpb8b"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.769917 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wpb8b" event={"ID":"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0","Type":"ContainerStarted","Data":"785378bdbb13106d2419874cd963ad5e6a590e05fce0f7cd116c27f25c3622ee"} Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.807571 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.822829 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pks9t"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.824363 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.827758 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.827957 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.837313 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pks9t"] Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.929601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-config-data\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.929690 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-scripts\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.929853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:44 crc kubenswrapper[4809]: I1127 17:40:44.929888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mpcr\" (UniqueName: \"kubernetes.io/projected/9dec0492-5e28-4725-a024-340d7562a2d9-kube-api-access-9mpcr\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.031226 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-scripts\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.031317 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.031341 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mpcr\" (UniqueName: \"kubernetes.io/projected/9dec0492-5e28-4725-a024-340d7562a2d9-kube-api-access-9mpcr\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.031729 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-config-data\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.035524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-scripts\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.036218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-config-data\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.065639 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mpcr\" (UniqueName: \"kubernetes.io/projected/9dec0492-5e28-4725-a024-340d7562a2d9-kube-api-access-9mpcr\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.075827 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pks9t\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.163536 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:45 crc kubenswrapper[4809]: W1127 17:40:45.241614 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9d0aec6_8a48_45a9_9b45_095b2bd10e51.slice/crio-a6f6b57e2b9172128c52d82946af6d88a24be5e6ca41896fd4834a802cf126a5 WatchSource:0}: Error finding container a6f6b57e2b9172128c52d82946af6d88a24be5e6ca41896fd4834a802cf126a5: Status 404 returned error can't find the container with id a6f6b57e2b9172128c52d82946af6d88a24be5e6ca41896fd4834a802cf126a5 Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.288579 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.517891 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.554418 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.569391 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xtlg4"] Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.782545 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" event={"ID":"16fbb102-acdd-49fd-a941-9e7ff32b71cc","Type":"ContainerStarted","Data":"9b4e73bb51e470d478112e8c3162fb03ad704e09a83cc7085c954d12afba6832"} Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.785718 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3c6e105-a2d2-4016-b84d-4444f89fb22f","Type":"ContainerStarted","Data":"5005c6c0587a480f5215a499f2bb2da9e3af6b2e50f292df580659552c4c3e79"} Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.788208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wpb8b" event={"ID":"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0","Type":"ContainerStarted","Data":"a6ca040fb04546a5edafdbad849c1f073cbe27d815f8e9bf3146650a709d5ad5"} Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.793395 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e19390b-a3ee-4b9c-b4b5-2e70134298d2","Type":"ContainerStarted","Data":"548dff10b9cc3d1b3466b3a6623ce71a8df0e3473678fd164481f89a574f2578"} Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.799426 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0ffdcceb-d220-4a44-88f9-94f7caf3694f","Type":"ContainerStarted","Data":"b08842eeeb2dd204f7b9564d27fc89861a2b60024c147ee0e8be2997c55f4ad1"} Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.801152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f9d0aec6-8a48-45a9-9b45-095b2bd10e51","Type":"ContainerStarted","Data":"a6f6b57e2b9172128c52d82946af6d88a24be5e6ca41896fd4834a802cf126a5"} Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.812361 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-wpb8b" podStartSLOduration=2.81233931 podStartE2EDuration="2.81233931s" podCreationTimestamp="2025-11-27 17:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:45.803701255 +0000 UTC m=+1881.076158607" watchObservedRunningTime="2025-11-27 17:40:45.81233931 +0000 UTC m=+1881.084796662" Nov 27 17:40:45 crc kubenswrapper[4809]: I1127 17:40:45.864575 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pks9t"] Nov 27 17:40:46 crc kubenswrapper[4809]: I1127 17:40:46.815449 4809 generic.go:334] "Generic (PLEG): container finished" podID="16fbb102-acdd-49fd-a941-9e7ff32b71cc" containerID="d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd" exitCode=0 Nov 27 17:40:46 crc kubenswrapper[4809]: I1127 17:40:46.815891 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" event={"ID":"16fbb102-acdd-49fd-a941-9e7ff32b71cc","Type":"ContainerDied","Data":"d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd"} Nov 27 17:40:46 crc kubenswrapper[4809]: I1127 17:40:46.821187 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pks9t" event={"ID":"9dec0492-5e28-4725-a024-340d7562a2d9","Type":"ContainerStarted","Data":"000c8417c4326e932eb3170750207e83f50bbff649ed548350a40a0035a9c1bf"} Nov 27 17:40:46 crc kubenswrapper[4809]: I1127 17:40:46.821228 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pks9t" event={"ID":"9dec0492-5e28-4725-a024-340d7562a2d9","Type":"ContainerStarted","Data":"58b7709945f0eaf59ae7d2e1f66b79a904ea1b86e46f1fa939524bbbd8724dae"} Nov 27 17:40:46 crc kubenswrapper[4809]: I1127 17:40:46.874911 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-pks9t" podStartSLOduration=2.874892137 podStartE2EDuration="2.874892137s" podCreationTimestamp="2025-11-27 17:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:46.855630835 +0000 UTC m=+1882.128088187" watchObservedRunningTime="2025-11-27 17:40:46.874892137 +0000 UTC m=+1882.147349489" Nov 27 17:40:48 crc kubenswrapper[4809]: I1127 17:40:48.259421 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:48 crc kubenswrapper[4809]: I1127 17:40:48.277132 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.871050 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" event={"ID":"16fbb102-acdd-49fd-a941-9e7ff32b71cc","Type":"ContainerStarted","Data":"84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490"} Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.871117 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.872852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3c6e105-a2d2-4016-b84d-4444f89fb22f","Type":"ContainerStarted","Data":"2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1"} Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.872903 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3c6e105-a2d2-4016-b84d-4444f89fb22f","Type":"ContainerStarted","Data":"712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733"} Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.874598 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e19390b-a3ee-4b9c-b4b5-2e70134298d2","Type":"ContainerStarted","Data":"74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694"} Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.875872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0ffdcceb-d220-4a44-88f9-94f7caf3694f","Type":"ContainerStarted","Data":"e0d64cac60b1d7dd70920d451042a26871b366b24598244ca13f72d1dea7dc9b"} Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.876008 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0ffdcceb-d220-4a44-88f9-94f7caf3694f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e0d64cac60b1d7dd70920d451042a26871b366b24598244ca13f72d1dea7dc9b" gracePeriod=30 Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.878784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f9d0aec6-8a48-45a9-9b45-095b2bd10e51","Type":"ContainerStarted","Data":"3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c"} Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.878825 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f9d0aec6-8a48-45a9-9b45-095b2bd10e51","Type":"ContainerStarted","Data":"7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f"} Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.878930 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerName="nova-metadata-log" containerID="cri-o://7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f" gracePeriod=30 Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.879047 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerName="nova-metadata-metadata" containerID="cri-o://3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c" gracePeriod=30 Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.904697 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" podStartSLOduration=5.904675241 podStartE2EDuration="5.904675241s" podCreationTimestamp="2025-11-27 17:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:49.893835648 +0000 UTC m=+1885.166293000" watchObservedRunningTime="2025-11-27 17:40:49.904675241 +0000 UTC m=+1885.177132593" Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.940262 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.16766552 podStartE2EDuration="6.940244876s" podCreationTimestamp="2025-11-27 17:40:43 +0000 UTC" firstStartedPulling="2025-11-27 17:40:44.8519442 +0000 UTC m=+1880.124401552" lastFinishedPulling="2025-11-27 17:40:48.624523556 +0000 UTC m=+1883.896980908" observedRunningTime="2025-11-27 17:40:49.932596378 +0000 UTC m=+1885.205053730" watchObservedRunningTime="2025-11-27 17:40:49.940244876 +0000 UTC m=+1885.212702228" Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.940671 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.8500429819999997 podStartE2EDuration="6.940666117s" podCreationTimestamp="2025-11-27 17:40:43 +0000 UTC" firstStartedPulling="2025-11-27 17:40:45.537702535 +0000 UTC m=+1880.810159887" lastFinishedPulling="2025-11-27 17:40:48.62832567 +0000 UTC m=+1883.900783022" observedRunningTime="2025-11-27 17:40:49.920237774 +0000 UTC m=+1885.192695126" watchObservedRunningTime="2025-11-27 17:40:49.940666117 +0000 UTC m=+1885.213123469" Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.963719 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.84698563 podStartE2EDuration="5.963699051s" podCreationTimestamp="2025-11-27 17:40:44 +0000 UTC" firstStartedPulling="2025-11-27 17:40:45.50981179 +0000 UTC m=+1880.782269142" lastFinishedPulling="2025-11-27 17:40:48.626525211 +0000 UTC m=+1883.898982563" observedRunningTime="2025-11-27 17:40:49.944994715 +0000 UTC m=+1885.217452067" watchObservedRunningTime="2025-11-27 17:40:49.963699051 +0000 UTC m=+1885.236156403" Nov 27 17:40:49 crc kubenswrapper[4809]: I1127 17:40:49.976559 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.639326631 podStartE2EDuration="6.976537019s" podCreationTimestamp="2025-11-27 17:40:43 +0000 UTC" firstStartedPulling="2025-11-27 17:40:45.287896785 +0000 UTC m=+1880.560354137" lastFinishedPulling="2025-11-27 17:40:48.625107173 +0000 UTC m=+1883.897564525" observedRunningTime="2025-11-27 17:40:49.964059342 +0000 UTC m=+1885.236516694" watchObservedRunningTime="2025-11-27 17:40:49.976537019 +0000 UTC m=+1885.248994371" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.685506 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.767916 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-logs\") pod \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.768112 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-combined-ca-bundle\") pod \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.768163 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-config-data\") pod \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.768234 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvw2x\" (UniqueName: \"kubernetes.io/projected/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-kube-api-access-zvw2x\") pod \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\" (UID: \"f9d0aec6-8a48-45a9-9b45-095b2bd10e51\") " Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.768386 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-logs" (OuterVolumeSpecName: "logs") pod "f9d0aec6-8a48-45a9-9b45-095b2bd10e51" (UID: "f9d0aec6-8a48-45a9-9b45-095b2bd10e51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.768676 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.773013 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-kube-api-access-zvw2x" (OuterVolumeSpecName: "kube-api-access-zvw2x") pod "f9d0aec6-8a48-45a9-9b45-095b2bd10e51" (UID: "f9d0aec6-8a48-45a9-9b45-095b2bd10e51"). InnerVolumeSpecName "kube-api-access-zvw2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.796700 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9d0aec6-8a48-45a9-9b45-095b2bd10e51" (UID: "f9d0aec6-8a48-45a9-9b45-095b2bd10e51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.803583 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-config-data" (OuterVolumeSpecName: "config-data") pod "f9d0aec6-8a48-45a9-9b45-095b2bd10e51" (UID: "f9d0aec6-8a48-45a9-9b45-095b2bd10e51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.871336 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.871361 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.871373 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvw2x\" (UniqueName: \"kubernetes.io/projected/f9d0aec6-8a48-45a9-9b45-095b2bd10e51-kube-api-access-zvw2x\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.890482 4809 generic.go:334] "Generic (PLEG): container finished" podID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerID="3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c" exitCode=0 Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.890534 4809 generic.go:334] "Generic (PLEG): container finished" podID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerID="7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f" exitCode=143 Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.890575 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.890642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f9d0aec6-8a48-45a9-9b45-095b2bd10e51","Type":"ContainerDied","Data":"3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c"} Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.890700 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f9d0aec6-8a48-45a9-9b45-095b2bd10e51","Type":"ContainerDied","Data":"7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f"} Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.890714 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f9d0aec6-8a48-45a9-9b45-095b2bd10e51","Type":"ContainerDied","Data":"a6f6b57e2b9172128c52d82946af6d88a24be5e6ca41896fd4834a802cf126a5"} Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.890761 4809 scope.go:117] "RemoveContainer" containerID="3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.921006 4809 scope.go:117] "RemoveContainer" containerID="7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.931444 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.950082 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.956489 4809 scope.go:117] "RemoveContainer" containerID="3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c" Nov 27 17:40:50 crc kubenswrapper[4809]: E1127 17:40:50.960860 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c\": container with ID starting with 3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c not found: ID does not exist" containerID="3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.960894 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c"} err="failed to get container status \"3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c\": rpc error: code = NotFound desc = could not find container \"3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c\": container with ID starting with 3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c not found: ID does not exist" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.960919 4809 scope.go:117] "RemoveContainer" containerID="7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f" Nov 27 17:40:50 crc kubenswrapper[4809]: E1127 17:40:50.961363 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f\": container with ID starting with 7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f not found: ID does not exist" containerID="7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.961388 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f"} err="failed to get container status \"7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f\": rpc error: code = NotFound desc = could not find container \"7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f\": container with ID starting with 7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f not found: ID does not exist" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.961404 4809 scope.go:117] "RemoveContainer" containerID="3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.963975 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c"} err="failed to get container status \"3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c\": rpc error: code = NotFound desc = could not find container \"3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c\": container with ID starting with 3dcfce150ed569429f4565e88ecb014dadd67fff5095bf96832e50f8211c687c not found: ID does not exist" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.964072 4809 scope.go:117] "RemoveContainer" containerID="7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.964462 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f"} err="failed to get container status \"7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f\": rpc error: code = NotFound desc = could not find container \"7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f\": container with ID starting with 7293e0e4ff0ad2a670ef3144aa6ade418e562e181e5117725407426cc5b9e60f not found: ID does not exist" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.966801 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:50 crc kubenswrapper[4809]: E1127 17:40:50.967289 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerName="nova-metadata-metadata" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.967312 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerName="nova-metadata-metadata" Nov 27 17:40:50 crc kubenswrapper[4809]: E1127 17:40:50.967364 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerName="nova-metadata-log" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.967374 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerName="nova-metadata-log" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.967618 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerName="nova-metadata-log" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.967659 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" containerName="nova-metadata-metadata" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.968965 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.971484 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 27 17:40:50 crc kubenswrapper[4809]: I1127 17:40:50.971610 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.019874 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.074386 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-config-data\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.074473 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.074501 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.074531 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg8mg\" (UniqueName: \"kubernetes.io/projected/7a67e20c-0d45-49f2-8fc3-08bc59f50159-kube-api-access-gg8mg\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.074822 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a67e20c-0d45-49f2-8fc3-08bc59f50159-logs\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.177545 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-config-data\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.178006 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.178050 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.178090 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg8mg\" (UniqueName: \"kubernetes.io/projected/7a67e20c-0d45-49f2-8fc3-08bc59f50159-kube-api-access-gg8mg\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.178188 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a67e20c-0d45-49f2-8fc3-08bc59f50159-logs\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.178821 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a67e20c-0d45-49f2-8fc3-08bc59f50159-logs\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.181440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.182210 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.193219 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-config-data\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.200161 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg8mg\" (UniqueName: \"kubernetes.io/projected/7a67e20c-0d45-49f2-8fc3-08bc59f50159-kube-api-access-gg8mg\") pod \"nova-metadata-0\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.313026 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.468095 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9d0aec6-8a48-45a9-9b45-095b2bd10e51" path="/var/lib/kubelet/pods/f9d0aec6-8a48-45a9-9b45-095b2bd10e51/volumes" Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.777140 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:51 crc kubenswrapper[4809]: I1127 17:40:51.905181 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a67e20c-0d45-49f2-8fc3-08bc59f50159","Type":"ContainerStarted","Data":"b8f17e4c34da973d0677be948723fa4b2329244312d7c2cdee994c81a739a569"} Nov 27 17:40:52 crc kubenswrapper[4809]: I1127 17:40:52.918504 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a67e20c-0d45-49f2-8fc3-08bc59f50159","Type":"ContainerStarted","Data":"2f12f08da0977ec4a82f97d6e7543e41c874274338cee8d4a212261510144e71"} Nov 27 17:40:52 crc kubenswrapper[4809]: I1127 17:40:52.919011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a67e20c-0d45-49f2-8fc3-08bc59f50159","Type":"ContainerStarted","Data":"dc310c7bd30a10934a903eb14185ec56d7ab859fe90e16c215a545f292c5cc9c"} Nov 27 17:40:52 crc kubenswrapper[4809]: I1127 17:40:52.939436 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.939416421 podStartE2EDuration="2.939416421s" podCreationTimestamp="2025-11-27 17:40:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:52.933173161 +0000 UTC m=+1888.205630533" watchObservedRunningTime="2025-11-27 17:40:52.939416421 +0000 UTC m=+1888.211873773" Nov 27 17:40:53 crc kubenswrapper[4809]: I1127 17:40:53.928196 4809 generic.go:334] "Generic (PLEG): container finished" podID="11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0" containerID="a6ca040fb04546a5edafdbad849c1f073cbe27d815f8e9bf3146650a709d5ad5" exitCode=0 Nov 27 17:40:53 crc kubenswrapper[4809]: I1127 17:40:53.928271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wpb8b" event={"ID":"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0","Type":"ContainerDied","Data":"a6ca040fb04546a5edafdbad849c1f073cbe27d815f8e9bf3146650a709d5ad5"} Nov 27 17:40:53 crc kubenswrapper[4809]: I1127 17:40:53.930162 4809 generic.go:334] "Generic (PLEG): container finished" podID="9dec0492-5e28-4725-a024-340d7562a2d9" containerID="000c8417c4326e932eb3170750207e83f50bbff649ed548350a40a0035a9c1bf" exitCode=0 Nov 27 17:40:53 crc kubenswrapper[4809]: I1127 17:40:53.930245 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pks9t" event={"ID":"9dec0492-5e28-4725-a024-340d7562a2d9","Type":"ContainerDied","Data":"000c8417c4326e932eb3170750207e83f50bbff649ed548350a40a0035a9c1bf"} Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.095206 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.095266 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.121538 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.458133 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:40:54 crc kubenswrapper[4809]: E1127 17:40:54.458425 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.472827 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.511515 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.511560 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.516617 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.608189 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5x7sr"] Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.608794 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" podUID="5ff6795d-3222-4d84-821b-86a303e0e74e" containerName="dnsmasq-dns" containerID="cri-o://122d6c5478b056259670d4dc4929caf6a5533924f924d753ef2b465d4788cd16" gracePeriod=10 Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.941149 4809 generic.go:334] "Generic (PLEG): container finished" podID="5ff6795d-3222-4d84-821b-86a303e0e74e" containerID="122d6c5478b056259670d4dc4929caf6a5533924f924d753ef2b465d4788cd16" exitCode=0 Nov 27 17:40:54 crc kubenswrapper[4809]: I1127 17:40:54.941330 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" event={"ID":"5ff6795d-3222-4d84-821b-86a303e0e74e","Type":"ContainerDied","Data":"122d6c5478b056259670d4dc4929caf6a5533924f924d753ef2b465d4788cd16"} Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.011357 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.216924 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.257393 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-svc\") pod \"5ff6795d-3222-4d84-821b-86a303e0e74e\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.257436 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-nb\") pod \"5ff6795d-3222-4d84-821b-86a303e0e74e\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.257615 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jpx9\" (UniqueName: \"kubernetes.io/projected/5ff6795d-3222-4d84-821b-86a303e0e74e-kube-api-access-2jpx9\") pod \"5ff6795d-3222-4d84-821b-86a303e0e74e\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.257632 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-sb\") pod \"5ff6795d-3222-4d84-821b-86a303e0e74e\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.257701 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-config\") pod \"5ff6795d-3222-4d84-821b-86a303e0e74e\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.257808 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-swift-storage-0\") pod \"5ff6795d-3222-4d84-821b-86a303e0e74e\" (UID: \"5ff6795d-3222-4d84-821b-86a303e0e74e\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.267359 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ff6795d-3222-4d84-821b-86a303e0e74e-kube-api-access-2jpx9" (OuterVolumeSpecName: "kube-api-access-2jpx9") pod "5ff6795d-3222-4d84-821b-86a303e0e74e" (UID: "5ff6795d-3222-4d84-821b-86a303e0e74e"). InnerVolumeSpecName "kube-api-access-2jpx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.361111 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jpx9\" (UniqueName: \"kubernetes.io/projected/5ff6795d-3222-4d84-821b-86a303e0e74e-kube-api-access-2jpx9\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.367973 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-config" (OuterVolumeSpecName: "config") pod "5ff6795d-3222-4d84-821b-86a303e0e74e" (UID: "5ff6795d-3222-4d84-821b-86a303e0e74e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.376364 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5ff6795d-3222-4d84-821b-86a303e0e74e" (UID: "5ff6795d-3222-4d84-821b-86a303e0e74e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.386158 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5ff6795d-3222-4d84-821b-86a303e0e74e" (UID: "5ff6795d-3222-4d84-821b-86a303e0e74e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.396933 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5ff6795d-3222-4d84-821b-86a303e0e74e" (UID: "5ff6795d-3222-4d84-821b-86a303e0e74e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.414381 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5ff6795d-3222-4d84-821b-86a303e0e74e" (UID: "5ff6795d-3222-4d84-821b-86a303e0e74e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.463231 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.463262 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.463274 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.463282 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.463309 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ff6795d-3222-4d84-821b-86a303e0e74e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.477829 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.487430 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.564308 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mpcr\" (UniqueName: \"kubernetes.io/projected/9dec0492-5e28-4725-a024-340d7562a2d9-kube-api-access-9mpcr\") pod \"9dec0492-5e28-4725-a024-340d7562a2d9\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.564444 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-combined-ca-bundle\") pod \"9dec0492-5e28-4725-a024-340d7562a2d9\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.564531 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-config-data\") pod \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.564902 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-combined-ca-bundle\") pod \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.565056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-config-data\") pod \"9dec0492-5e28-4725-a024-340d7562a2d9\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.565168 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-scripts\") pod \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.565463 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-scripts\") pod \"9dec0492-5e28-4725-a024-340d7562a2d9\" (UID: \"9dec0492-5e28-4725-a024-340d7562a2d9\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.565545 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzzrc\" (UniqueName: \"kubernetes.io/projected/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-kube-api-access-vzzrc\") pod \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\" (UID: \"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0\") " Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.568308 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dec0492-5e28-4725-a024-340d7562a2d9-kube-api-access-9mpcr" (OuterVolumeSpecName: "kube-api-access-9mpcr") pod "9dec0492-5e28-4725-a024-340d7562a2d9" (UID: "9dec0492-5e28-4725-a024-340d7562a2d9"). InnerVolumeSpecName "kube-api-access-9mpcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.569916 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-scripts" (OuterVolumeSpecName: "scripts") pod "11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0" (UID: "11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.570166 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-scripts" (OuterVolumeSpecName: "scripts") pod "9dec0492-5e28-4725-a024-340d7562a2d9" (UID: "9dec0492-5e28-4725-a024-340d7562a2d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.573595 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-kube-api-access-vzzrc" (OuterVolumeSpecName: "kube-api-access-vzzrc") pod "11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0" (UID: "11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0"). InnerVolumeSpecName "kube-api-access-vzzrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.593901 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.593905 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.595057 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-config-data" (OuterVolumeSpecName: "config-data") pod "9dec0492-5e28-4725-a024-340d7562a2d9" (UID: "9dec0492-5e28-4725-a024-340d7562a2d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.595222 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-config-data" (OuterVolumeSpecName: "config-data") pod "11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0" (UID: "11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.600722 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0" (UID: "11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.601596 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9dec0492-5e28-4725-a024-340d7562a2d9" (UID: "9dec0492-5e28-4725-a024-340d7562a2d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.667869 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mpcr\" (UniqueName: \"kubernetes.io/projected/9dec0492-5e28-4725-a024-340d7562a2d9-kube-api-access-9mpcr\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.667905 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.667920 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.667938 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.667950 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.667963 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.667974 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dec0492-5e28-4725-a024-340d7562a2d9-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.667985 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzzrc\" (UniqueName: \"kubernetes.io/projected/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0-kube-api-access-vzzrc\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.954416 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" event={"ID":"5ff6795d-3222-4d84-821b-86a303e0e74e","Type":"ContainerDied","Data":"93cb8a84c482744cbc1dcbbae348d8eb2101f4ce7b86999c0962dbe1b153ffbe"} Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.954481 4809 scope.go:117] "RemoveContainer" containerID="122d6c5478b056259670d4dc4929caf6a5533924f924d753ef2b465d4788cd16" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.954626 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-5x7sr" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.959686 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wpb8b" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.960052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wpb8b" event={"ID":"11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0","Type":"ContainerDied","Data":"785378bdbb13106d2419874cd963ad5e6a590e05fce0f7cd116c27f25c3622ee"} Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.960093 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="785378bdbb13106d2419874cd963ad5e6a590e05fce0f7cd116c27f25c3622ee" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.964174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pks9t" event={"ID":"9dec0492-5e28-4725-a024-340d7562a2d9","Type":"ContainerDied","Data":"58b7709945f0eaf59ae7d2e1f66b79a904ea1b86e46f1fa939524bbbd8724dae"} Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.964220 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58b7709945f0eaf59ae7d2e1f66b79a904ea1b86e46f1fa939524bbbd8724dae" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.964352 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pks9t" Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.989397 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5x7sr"] Nov 27 17:40:55 crc kubenswrapper[4809]: I1127 17:40:55.992913 4809 scope.go:117] "RemoveContainer" containerID="41a970cdf1341fe0baa0a36ca62e94be78e1d62997846ac4a45ae696e3c87978" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.015016 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5x7sr"] Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.077110 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 27 17:40:56 crc kubenswrapper[4809]: E1127 17:40:56.077824 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff6795d-3222-4d84-821b-86a303e0e74e" containerName="dnsmasq-dns" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.077849 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff6795d-3222-4d84-821b-86a303e0e74e" containerName="dnsmasq-dns" Nov 27 17:40:56 crc kubenswrapper[4809]: E1127 17:40:56.077875 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0" containerName="nova-manage" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.077884 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0" containerName="nova-manage" Nov 27 17:40:56 crc kubenswrapper[4809]: E1127 17:40:56.077912 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff6795d-3222-4d84-821b-86a303e0e74e" containerName="init" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.077922 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff6795d-3222-4d84-821b-86a303e0e74e" containerName="init" Nov 27 17:40:56 crc kubenswrapper[4809]: E1127 17:40:56.077933 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dec0492-5e28-4725-a024-340d7562a2d9" containerName="nova-cell1-conductor-db-sync" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.077941 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dec0492-5e28-4725-a024-340d7562a2d9" containerName="nova-cell1-conductor-db-sync" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.078202 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0" containerName="nova-manage" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.078238 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dec0492-5e28-4725-a024-340d7562a2d9" containerName="nova-cell1-conductor-db-sync" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.078259 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ff6795d-3222-4d84-821b-86a303e0e74e" containerName="dnsmasq-dns" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.080138 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.086531 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.086660 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.152867 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.153126 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-log" containerID="cri-o://712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733" gracePeriod=30 Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.153157 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-api" containerID="cri-o://2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1" gracePeriod=30 Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.165364 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.180130 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ecca97-66c5-4670-9304-5fb2b116e4be-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"01ecca97-66c5-4670-9304-5fb2b116e4be\") " pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.180349 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ecca97-66c5-4670-9304-5fb2b116e4be-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"01ecca97-66c5-4670-9304-5fb2b116e4be\") " pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.180452 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpw25\" (UniqueName: \"kubernetes.io/projected/01ecca97-66c5-4670-9304-5fb2b116e4be-kube-api-access-lpw25\") pod \"nova-cell1-conductor-0\" (UID: \"01ecca97-66c5-4670-9304-5fb2b116e4be\") " pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.191697 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.192083 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerName="nova-metadata-log" containerID="cri-o://dc310c7bd30a10934a903eb14185ec56d7ab859fe90e16c215a545f292c5cc9c" gracePeriod=30 Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.192231 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerName="nova-metadata-metadata" containerID="cri-o://2f12f08da0977ec4a82f97d6e7543e41c874274338cee8d4a212261510144e71" gracePeriod=30 Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.282627 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ecca97-66c5-4670-9304-5fb2b116e4be-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"01ecca97-66c5-4670-9304-5fb2b116e4be\") " pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.282710 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ecca97-66c5-4670-9304-5fb2b116e4be-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"01ecca97-66c5-4670-9304-5fb2b116e4be\") " pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.282759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpw25\" (UniqueName: \"kubernetes.io/projected/01ecca97-66c5-4670-9304-5fb2b116e4be-kube-api-access-lpw25\") pod \"nova-cell1-conductor-0\" (UID: \"01ecca97-66c5-4670-9304-5fb2b116e4be\") " pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.288077 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ecca97-66c5-4670-9304-5fb2b116e4be-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"01ecca97-66c5-4670-9304-5fb2b116e4be\") " pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.292607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ecca97-66c5-4670-9304-5fb2b116e4be-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"01ecca97-66c5-4670-9304-5fb2b116e4be\") " pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.302435 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpw25\" (UniqueName: \"kubernetes.io/projected/01ecca97-66c5-4670-9304-5fb2b116e4be-kube-api-access-lpw25\") pod \"nova-cell1-conductor-0\" (UID: \"01ecca97-66c5-4670-9304-5fb2b116e4be\") " pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.313790 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.313846 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.401455 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.982908 4809 generic.go:334] "Generic (PLEG): container finished" podID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerID="2f12f08da0977ec4a82f97d6e7543e41c874274338cee8d4a212261510144e71" exitCode=0 Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.983513 4809 generic.go:334] "Generic (PLEG): container finished" podID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerID="dc310c7bd30a10934a903eb14185ec56d7ab859fe90e16c215a545f292c5cc9c" exitCode=143 Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.983051 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a67e20c-0d45-49f2-8fc3-08bc59f50159","Type":"ContainerDied","Data":"2f12f08da0977ec4a82f97d6e7543e41c874274338cee8d4a212261510144e71"} Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.983603 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a67e20c-0d45-49f2-8fc3-08bc59f50159","Type":"ContainerDied","Data":"dc310c7bd30a10934a903eb14185ec56d7ab859fe90e16c215a545f292c5cc9c"} Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.983616 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7a67e20c-0d45-49f2-8fc3-08bc59f50159","Type":"ContainerDied","Data":"b8f17e4c34da973d0677be948723fa4b2329244312d7c2cdee994c81a739a569"} Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.983626 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8f17e4c34da973d0677be948723fa4b2329244312d7c2cdee994c81a739a569" Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.996872 4809 generic.go:334] "Generic (PLEG): container finished" podID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerID="712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733" exitCode=143 Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.997035 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3c6e105-a2d2-4016-b84d-4444f89fb22f","Type":"ContainerDied","Data":"712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733"} Nov 27 17:40:56 crc kubenswrapper[4809]: I1127 17:40:56.999642 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7e19390b-a3ee-4b9c-b4b5-2e70134298d2" containerName="nova-scheduler-scheduler" containerID="cri-o://74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694" gracePeriod=30 Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.041310 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.097177 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-config-data\") pod \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.097221 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-nova-metadata-tls-certs\") pod \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.097292 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg8mg\" (UniqueName: \"kubernetes.io/projected/7a67e20c-0d45-49f2-8fc3-08bc59f50159-kube-api-access-gg8mg\") pod \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.097511 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a67e20c-0d45-49f2-8fc3-08bc59f50159-logs\") pod \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.097776 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-combined-ca-bundle\") pod \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\" (UID: \"7a67e20c-0d45-49f2-8fc3-08bc59f50159\") " Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.098562 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a67e20c-0d45-49f2-8fc3-08bc59f50159-logs" (OuterVolumeSpecName: "logs") pod "7a67e20c-0d45-49f2-8fc3-08bc59f50159" (UID: "7a67e20c-0d45-49f2-8fc3-08bc59f50159"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.107061 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a67e20c-0d45-49f2-8fc3-08bc59f50159-kube-api-access-gg8mg" (OuterVolumeSpecName: "kube-api-access-gg8mg") pod "7a67e20c-0d45-49f2-8fc3-08bc59f50159" (UID: "7a67e20c-0d45-49f2-8fc3-08bc59f50159"). InnerVolumeSpecName "kube-api-access-gg8mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.137875 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-config-data" (OuterVolumeSpecName: "config-data") pod "7a67e20c-0d45-49f2-8fc3-08bc59f50159" (UID: "7a67e20c-0d45-49f2-8fc3-08bc59f50159"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.161437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a67e20c-0d45-49f2-8fc3-08bc59f50159" (UID: "7a67e20c-0d45-49f2-8fc3-08bc59f50159"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.168391 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 27 17:40:57 crc kubenswrapper[4809]: W1127 17:40:57.173138 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ecca97_66c5_4670_9304_5fb2b116e4be.slice/crio-ded4a27d24e795e3dc3805b2b195198e2a37ed1b430c529f3c7147839bb31724 WatchSource:0}: Error finding container ded4a27d24e795e3dc3805b2b195198e2a37ed1b430c529f3c7147839bb31724: Status 404 returned error can't find the container with id ded4a27d24e795e3dc3805b2b195198e2a37ed1b430c529f3c7147839bb31724 Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.175401 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7a67e20c-0d45-49f2-8fc3-08bc59f50159" (UID: "7a67e20c-0d45-49f2-8fc3-08bc59f50159"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.199930 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.199962 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.199971 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg8mg\" (UniqueName: \"kubernetes.io/projected/7a67e20c-0d45-49f2-8fc3-08bc59f50159-kube-api-access-gg8mg\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.199981 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a67e20c-0d45-49f2-8fc3-08bc59f50159-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.199989 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a67e20c-0d45-49f2-8fc3-08bc59f50159-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:40:57 crc kubenswrapper[4809]: I1127 17:40:57.470077 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ff6795d-3222-4d84-821b-86a303e0e74e" path="/var/lib/kubelet/pods/5ff6795d-3222-4d84-821b-86a303e0e74e/volumes" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.010793 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.012006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"01ecca97-66c5-4670-9304-5fb2b116e4be","Type":"ContainerStarted","Data":"80294a152e96feeb4ae94da727cdfe3e0a8150bc038a11091ccaccd0c6fe6cf0"} Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.012040 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.012051 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"01ecca97-66c5-4670-9304-5fb2b116e4be","Type":"ContainerStarted","Data":"ded4a27d24e795e3dc3805b2b195198e2a37ed1b430c529f3c7147839bb31724"} Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.032354 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.032334293 podStartE2EDuration="2.032334293s" podCreationTimestamp="2025-11-27 17:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:40:58.030983456 +0000 UTC m=+1893.303440808" watchObservedRunningTime="2025-11-27 17:40:58.032334293 +0000 UTC m=+1893.304791645" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.055964 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.070409 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.093317 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:58 crc kubenswrapper[4809]: E1127 17:40:58.093862 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerName="nova-metadata-metadata" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.093887 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerName="nova-metadata-metadata" Nov 27 17:40:58 crc kubenswrapper[4809]: E1127 17:40:58.093923 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerName="nova-metadata-log" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.093933 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerName="nova-metadata-log" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.094209 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerName="nova-metadata-log" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.094256 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" containerName="nova-metadata-metadata" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.095636 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.103661 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.103922 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.104441 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.227217 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-config-data\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.227270 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8232a1f9-b58b-4e35-87ec-5d14431c573a-logs\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.227291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwvcb\" (UniqueName: \"kubernetes.io/projected/8232a1f9-b58b-4e35-87ec-5d14431c573a-kube-api-access-fwvcb\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.227315 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.227409 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.329259 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-config-data\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.329321 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8232a1f9-b58b-4e35-87ec-5d14431c573a-logs\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.329343 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwvcb\" (UniqueName: \"kubernetes.io/projected/8232a1f9-b58b-4e35-87ec-5d14431c573a-kube-api-access-fwvcb\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.329370 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.329405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.329786 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8232a1f9-b58b-4e35-87ec-5d14431c573a-logs\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.334994 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.335002 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-config-data\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.335716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.345414 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwvcb\" (UniqueName: \"kubernetes.io/projected/8232a1f9-b58b-4e35-87ec-5d14431c573a-kube-api-access-fwvcb\") pod \"nova-metadata-0\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.424036 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:40:58 crc kubenswrapper[4809]: I1127 17:40:58.903431 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:40:59 crc kubenswrapper[4809]: I1127 17:40:59.020934 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8232a1f9-b58b-4e35-87ec-5d14431c573a","Type":"ContainerStarted","Data":"873de81636e9b7a52dd578be7e11a04d11fe526947e312fd91083d85848c98d2"} Nov 27 17:40:59 crc kubenswrapper[4809]: E1127 17:40:59.098126 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 27 17:40:59 crc kubenswrapper[4809]: E1127 17:40:59.099458 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 27 17:40:59 crc kubenswrapper[4809]: E1127 17:40:59.100573 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 27 17:40:59 crc kubenswrapper[4809]: E1127 17:40:59.100635 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7e19390b-a3ee-4b9c-b4b5-2e70134298d2" containerName="nova-scheduler-scheduler" Nov 27 17:40:59 crc kubenswrapper[4809]: I1127 17:40:59.471234 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a67e20c-0d45-49f2-8fc3-08bc59f50159" path="/var/lib/kubelet/pods/7a67e20c-0d45-49f2-8fc3-08bc59f50159/volumes" Nov 27 17:41:00 crc kubenswrapper[4809]: I1127 17:41:00.032414 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8232a1f9-b58b-4e35-87ec-5d14431c573a","Type":"ContainerStarted","Data":"96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107"} Nov 27 17:41:00 crc kubenswrapper[4809]: I1127 17:41:00.032468 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8232a1f9-b58b-4e35-87ec-5d14431c573a","Type":"ContainerStarted","Data":"3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3"} Nov 27 17:41:00 crc kubenswrapper[4809]: I1127 17:41:00.055340 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.0553218109999998 podStartE2EDuration="2.055321811s" podCreationTimestamp="2025-11-27 17:40:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:00.049315209 +0000 UTC m=+1895.321772561" watchObservedRunningTime="2025-11-27 17:41:00.055321811 +0000 UTC m=+1895.327779183" Nov 27 17:41:00 crc kubenswrapper[4809]: I1127 17:41:00.985043 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.044625 4809 generic.go:334] "Generic (PLEG): container finished" podID="7e19390b-a3ee-4b9c-b4b5-2e70134298d2" containerID="74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694" exitCode=0 Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.044981 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e19390b-a3ee-4b9c-b4b5-2e70134298d2","Type":"ContainerDied","Data":"74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694"} Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.045010 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.045050 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e19390b-a3ee-4b9c-b4b5-2e70134298d2","Type":"ContainerDied","Data":"548dff10b9cc3d1b3466b3a6623ce71a8df0e3473678fd164481f89a574f2578"} Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.045070 4809 scope.go:117] "RemoveContainer" containerID="74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.068644 4809 scope.go:117] "RemoveContainer" containerID="74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694" Nov 27 17:41:01 crc kubenswrapper[4809]: E1127 17:41:01.069158 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694\": container with ID starting with 74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694 not found: ID does not exist" containerID="74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.069203 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694"} err="failed to get container status \"74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694\": rpc error: code = NotFound desc = could not find container \"74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694\": container with ID starting with 74a7de9c1e966facfab98134319ef9168d507585cc8eef66f013ee86ef1b6694 not found: ID does not exist" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.082016 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-combined-ca-bundle\") pod \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.082264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-config-data\") pod \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.082287 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5wfz\" (UniqueName: \"kubernetes.io/projected/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-kube-api-access-x5wfz\") pod \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\" (UID: \"7e19390b-a3ee-4b9c-b4b5-2e70134298d2\") " Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.088055 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-kube-api-access-x5wfz" (OuterVolumeSpecName: "kube-api-access-x5wfz") pod "7e19390b-a3ee-4b9c-b4b5-2e70134298d2" (UID: "7e19390b-a3ee-4b9c-b4b5-2e70134298d2"). InnerVolumeSpecName "kube-api-access-x5wfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.118602 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-config-data" (OuterVolumeSpecName: "config-data") pod "7e19390b-a3ee-4b9c-b4b5-2e70134298d2" (UID: "7e19390b-a3ee-4b9c-b4b5-2e70134298d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.119910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e19390b-a3ee-4b9c-b4b5-2e70134298d2" (UID: "7e19390b-a3ee-4b9c-b4b5-2e70134298d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.184905 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.185133 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5wfz\" (UniqueName: \"kubernetes.io/projected/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-kube-api-access-x5wfz\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.185217 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e19390b-a3ee-4b9c-b4b5-2e70134298d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.379468 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.388953 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.407325 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:01 crc kubenswrapper[4809]: E1127 17:41:01.408071 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e19390b-a3ee-4b9c-b4b5-2e70134298d2" containerName="nova-scheduler-scheduler" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.408097 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e19390b-a3ee-4b9c-b4b5-2e70134298d2" containerName="nova-scheduler-scheduler" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.408438 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e19390b-a3ee-4b9c-b4b5-2e70134298d2" containerName="nova-scheduler-scheduler" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.409858 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.417264 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.421221 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.472508 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e19390b-a3ee-4b9c-b4b5-2e70134298d2" path="/var/lib/kubelet/pods/7e19390b-a3ee-4b9c-b4b5-2e70134298d2/volumes" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.492259 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-config-data\") pod \"nova-scheduler-0\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.492633 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.492932 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxn85\" (UniqueName: \"kubernetes.io/projected/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-kube-api-access-dxn85\") pod \"nova-scheduler-0\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.595709 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxn85\" (UniqueName: \"kubernetes.io/projected/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-kube-api-access-dxn85\") pod \"nova-scheduler-0\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.595873 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-config-data\") pod \"nova-scheduler-0\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.595926 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.601917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.601930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-config-data\") pod \"nova-scheduler-0\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.637051 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxn85\" (UniqueName: \"kubernetes.io/projected/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-kube-api-access-dxn85\") pod \"nova-scheduler-0\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.730998 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:41:01 crc kubenswrapper[4809]: I1127 17:41:01.966827 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.002583 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlnjx\" (UniqueName: \"kubernetes.io/projected/d3c6e105-a2d2-4016-b84d-4444f89fb22f-kube-api-access-vlnjx\") pod \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.002940 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3c6e105-a2d2-4016-b84d-4444f89fb22f-logs\") pod \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.003254 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-config-data\") pod \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.003445 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3c6e105-a2d2-4016-b84d-4444f89fb22f-logs" (OuterVolumeSpecName: "logs") pod "d3c6e105-a2d2-4016-b84d-4444f89fb22f" (UID: "d3c6e105-a2d2-4016-b84d-4444f89fb22f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.003472 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-combined-ca-bundle\") pod \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\" (UID: \"d3c6e105-a2d2-4016-b84d-4444f89fb22f\") " Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.004329 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3c6e105-a2d2-4016-b84d-4444f89fb22f-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.013214 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3c6e105-a2d2-4016-b84d-4444f89fb22f-kube-api-access-vlnjx" (OuterVolumeSpecName: "kube-api-access-vlnjx") pod "d3c6e105-a2d2-4016-b84d-4444f89fb22f" (UID: "d3c6e105-a2d2-4016-b84d-4444f89fb22f"). InnerVolumeSpecName "kube-api-access-vlnjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.038980 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3c6e105-a2d2-4016-b84d-4444f89fb22f" (UID: "d3c6e105-a2d2-4016-b84d-4444f89fb22f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.041092 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-config-data" (OuterVolumeSpecName: "config-data") pod "d3c6e105-a2d2-4016-b84d-4444f89fb22f" (UID: "d3c6e105-a2d2-4016-b84d-4444f89fb22f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.063217 4809 generic.go:334] "Generic (PLEG): container finished" podID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerID="2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1" exitCode=0 Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.064714 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3c6e105-a2d2-4016-b84d-4444f89fb22f","Type":"ContainerDied","Data":"2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1"} Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.064816 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.064919 4809 scope.go:117] "RemoveContainer" containerID="2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.065243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3c6e105-a2d2-4016-b84d-4444f89fb22f","Type":"ContainerDied","Data":"5005c6c0587a480f5215a499f2bb2da9e3af6b2e50f292df580659552c4c3e79"} Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.101638 4809 scope.go:117] "RemoveContainer" containerID="712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.107433 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.107468 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3c6e105-a2d2-4016-b84d-4444f89fb22f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.107484 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlnjx\" (UniqueName: \"kubernetes.io/projected/d3c6e105-a2d2-4016-b84d-4444f89fb22f-kube-api-access-vlnjx\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.124166 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.139728 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.146579 4809 scope.go:117] "RemoveContainer" containerID="2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1" Nov 27 17:41:02 crc kubenswrapper[4809]: E1127 17:41:02.147094 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1\": container with ID starting with 2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1 not found: ID does not exist" containerID="2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.147132 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1"} err="failed to get container status \"2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1\": rpc error: code = NotFound desc = could not find container \"2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1\": container with ID starting with 2c98376b08a8f6ef12b567bd75daaab59c168b2cbd6a0e1db8552f715a44f1e1 not found: ID does not exist" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.147164 4809 scope.go:117] "RemoveContainer" containerID="712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733" Nov 27 17:41:02 crc kubenswrapper[4809]: E1127 17:41:02.147486 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733\": container with ID starting with 712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733 not found: ID does not exist" containerID="712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.147512 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733"} err="failed to get container status \"712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733\": rpc error: code = NotFound desc = could not find container \"712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733\": container with ID starting with 712996fe50e6af3c3b2857d51a112db297b9d21087e419d30f68072e1b817733 not found: ID does not exist" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.154810 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:02 crc kubenswrapper[4809]: E1127 17:41:02.155386 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-api" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.155416 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-api" Nov 27 17:41:02 crc kubenswrapper[4809]: E1127 17:41:02.155471 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-log" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.155481 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-log" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.155717 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-log" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.155788 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" containerName="nova-api-api" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.166241 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.171590 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.180280 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.215848 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1895833b-2f96-438e-8ab2-85b1d9752b87-logs\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.216056 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.216150 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp2z9\" (UniqueName: \"kubernetes.io/projected/1895833b-2f96-438e-8ab2-85b1d9752b87-kube-api-access-mp2z9\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.216309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-config-data\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.244123 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.319696 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-config-data\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.319990 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1895833b-2f96-438e-8ab2-85b1d9752b87-logs\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.320120 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.320254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp2z9\" (UniqueName: \"kubernetes.io/projected/1895833b-2f96-438e-8ab2-85b1d9752b87-kube-api-access-mp2z9\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.320817 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1895833b-2f96-438e-8ab2-85b1d9752b87-logs\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.325385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-config-data\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.326884 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.337358 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp2z9\" (UniqueName: \"kubernetes.io/projected/1895833b-2f96-438e-8ab2-85b1d9752b87-kube-api-access-mp2z9\") pod \"nova-api-0\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " pod="openstack/nova-api-0" Nov 27 17:41:02 crc kubenswrapper[4809]: I1127 17:41:02.490377 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:03 crc kubenswrapper[4809]: I1127 17:41:03.081151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6118e4d-990e-42f3-8ab2-584e7b29e3b9","Type":"ContainerStarted","Data":"875f2953176d32da50aab1ff7340959330701d3059b1571068e337680dc9dad8"} Nov 27 17:41:03 crc kubenswrapper[4809]: I1127 17:41:03.081517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6118e4d-990e-42f3-8ab2-584e7b29e3b9","Type":"ContainerStarted","Data":"f8ed8dbe1dcb084bb5b3c9d77d343cee0f27fba6821d7b45672c9f82a93f4e5a"} Nov 27 17:41:03 crc kubenswrapper[4809]: I1127 17:41:03.105662 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.105638073 podStartE2EDuration="2.105638073s" podCreationTimestamp="2025-11-27 17:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:03.098779386 +0000 UTC m=+1898.371236738" watchObservedRunningTime="2025-11-27 17:41:03.105638073 +0000 UTC m=+1898.378095435" Nov 27 17:41:03 crc kubenswrapper[4809]: I1127 17:41:03.121421 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:03 crc kubenswrapper[4809]: I1127 17:41:03.425191 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 17:41:03 crc kubenswrapper[4809]: I1127 17:41:03.426977 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 17:41:03 crc kubenswrapper[4809]: I1127 17:41:03.490449 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3c6e105-a2d2-4016-b84d-4444f89fb22f" path="/var/lib/kubelet/pods/d3c6e105-a2d2-4016-b84d-4444f89fb22f/volumes" Nov 27 17:41:04 crc kubenswrapper[4809]: I1127 17:41:04.103808 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1895833b-2f96-438e-8ab2-85b1d9752b87","Type":"ContainerStarted","Data":"5cecd8bc240027b6b665d7bd337bc08fa70cf6bdf17404efc8a9e9b7ad749ba2"} Nov 27 17:41:04 crc kubenswrapper[4809]: I1127 17:41:04.104404 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1895833b-2f96-438e-8ab2-85b1d9752b87","Type":"ContainerStarted","Data":"49dcdb1d225f90d0f63215dd008123276303a47ad31fafddf95d0ab07dffbcdd"} Nov 27 17:41:04 crc kubenswrapper[4809]: I1127 17:41:04.104420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1895833b-2f96-438e-8ab2-85b1d9752b87","Type":"ContainerStarted","Data":"2998d53f72aabdcba55037edf00afabf5505720f2485719642a0f68648927a20"} Nov 27 17:41:04 crc kubenswrapper[4809]: I1127 17:41:04.148254 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.148223189 podStartE2EDuration="2.148223189s" podCreationTimestamp="2025-11-27 17:41:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:04.136132551 +0000 UTC m=+1899.408589933" watchObservedRunningTime="2025-11-27 17:41:04.148223189 +0000 UTC m=+1899.420680541" Nov 27 17:41:05 crc kubenswrapper[4809]: I1127 17:41:05.000921 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 27 17:41:06 crc kubenswrapper[4809]: I1127 17:41:06.431730 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 27 17:41:06 crc kubenswrapper[4809]: I1127 17:41:06.731568 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 27 17:41:08 crc kubenswrapper[4809]: I1127 17:41:08.373268 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 17:41:08 crc kubenswrapper[4809]: I1127 17:41:08.374716 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="07137204-40d7-4004-a77d-6b704e1d1602" containerName="kube-state-metrics" containerID="cri-o://9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4" gracePeriod=30 Nov 27 17:41:08 crc kubenswrapper[4809]: I1127 17:41:08.425151 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 17:41:08 crc kubenswrapper[4809]: I1127 17:41:08.425677 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 17:41:08 crc kubenswrapper[4809]: I1127 17:41:08.952606 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.056014 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w57f\" (UniqueName: \"kubernetes.io/projected/07137204-40d7-4004-a77d-6b704e1d1602-kube-api-access-6w57f\") pod \"07137204-40d7-4004-a77d-6b704e1d1602\" (UID: \"07137204-40d7-4004-a77d-6b704e1d1602\") " Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.065963 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07137204-40d7-4004-a77d-6b704e1d1602-kube-api-access-6w57f" (OuterVolumeSpecName: "kube-api-access-6w57f") pod "07137204-40d7-4004-a77d-6b704e1d1602" (UID: "07137204-40d7-4004-a77d-6b704e1d1602"). InnerVolumeSpecName "kube-api-access-6w57f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.153604 4809 generic.go:334] "Generic (PLEG): container finished" podID="07137204-40d7-4004-a77d-6b704e1d1602" containerID="9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4" exitCode=2 Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.154953 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.156051 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"07137204-40d7-4004-a77d-6b704e1d1602","Type":"ContainerDied","Data":"9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4"} Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.156106 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"07137204-40d7-4004-a77d-6b704e1d1602","Type":"ContainerDied","Data":"ceccb9d22686c2d8db5df424e123d28f91c3411f6407362cd75305af8361a96c"} Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.156125 4809 scope.go:117] "RemoveContainer" containerID="9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.159108 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w57f\" (UniqueName: \"kubernetes.io/projected/07137204-40d7-4004-a77d-6b704e1d1602-kube-api-access-6w57f\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.180866 4809 scope.go:117] "RemoveContainer" containerID="9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4" Nov 27 17:41:09 crc kubenswrapper[4809]: E1127 17:41:09.181336 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4\": container with ID starting with 9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4 not found: ID does not exist" containerID="9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.181383 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4"} err="failed to get container status \"9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4\": rpc error: code = NotFound desc = could not find container \"9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4\": container with ID starting with 9ac286e6f550d9a767124bbd8ee9447e955c13281a2e46d85b8a0219658469a4 not found: ID does not exist" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.204145 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.224785 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.237322 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 17:41:09 crc kubenswrapper[4809]: E1127 17:41:09.237891 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07137204-40d7-4004-a77d-6b704e1d1602" containerName="kube-state-metrics" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.237915 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="07137204-40d7-4004-a77d-6b704e1d1602" containerName="kube-state-metrics" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.238200 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="07137204-40d7-4004-a77d-6b704e1d1602" containerName="kube-state-metrics" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.239161 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.242834 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.242834 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.250235 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.362863 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x949s\" (UniqueName: \"kubernetes.io/projected/3ff647e6-3fcb-401c-a695-047f07ca9a58-kube-api-access-x949s\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.362925 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff647e6-3fcb-401c-a695-047f07ca9a58-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.363073 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ff647e6-3fcb-401c-a695-047f07ca9a58-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.363200 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3ff647e6-3fcb-401c-a695-047f07ca9a58-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.438941 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.438978 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.458717 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:41:09 crc kubenswrapper[4809]: E1127 17:41:09.459096 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.464917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3ff647e6-3fcb-401c-a695-047f07ca9a58-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.465059 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x949s\" (UniqueName: \"kubernetes.io/projected/3ff647e6-3fcb-401c-a695-047f07ca9a58-kube-api-access-x949s\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.465106 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff647e6-3fcb-401c-a695-047f07ca9a58-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.465168 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ff647e6-3fcb-401c-a695-047f07ca9a58-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.473412 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3ff647e6-3fcb-401c-a695-047f07ca9a58-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.476513 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ff647e6-3fcb-401c-a695-047f07ca9a58-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.485963 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff647e6-3fcb-401c-a695-047f07ca9a58-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.487771 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07137204-40d7-4004-a77d-6b704e1d1602" path="/var/lib/kubelet/pods/07137204-40d7-4004-a77d-6b704e1d1602/volumes" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.511853 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x949s\" (UniqueName: \"kubernetes.io/projected/3ff647e6-3fcb-401c-a695-047f07ca9a58-kube-api-access-x949s\") pod \"kube-state-metrics-0\" (UID: \"3ff647e6-3fcb-401c-a695-047f07ca9a58\") " pod="openstack/kube-state-metrics-0" Nov 27 17:41:09 crc kubenswrapper[4809]: I1127 17:41:09.574276 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 17:41:10 crc kubenswrapper[4809]: I1127 17:41:10.198708 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 17:41:10 crc kubenswrapper[4809]: I1127 17:41:10.419018 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:10 crc kubenswrapper[4809]: I1127 17:41:10.420876 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="ceilometer-central-agent" containerID="cri-o://c52a6040f3804de9369eccd5891abcf98bff5e7d1b87d9f9a1ceb591946cad99" gracePeriod=30 Nov 27 17:41:10 crc kubenswrapper[4809]: I1127 17:41:10.420988 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="sg-core" containerID="cri-o://bd8eec56078ff1a5910c47c787dadb501a8f077b8412160d3774e70541b98b24" gracePeriod=30 Nov 27 17:41:10 crc kubenswrapper[4809]: I1127 17:41:10.421039 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="ceilometer-notification-agent" containerID="cri-o://162fdb74efc4fa762b0e1378da20e568bcd84b3daeea002661753778e9f08150" gracePeriod=30 Nov 27 17:41:10 crc kubenswrapper[4809]: I1127 17:41:10.421222 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="proxy-httpd" containerID="cri-o://4642b43a93b98e574260a834a13c519c1eda1ddb903a08eccc2d908cada91d75" gracePeriod=30 Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.175678 4809 generic.go:334] "Generic (PLEG): container finished" podID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerID="4642b43a93b98e574260a834a13c519c1eda1ddb903a08eccc2d908cada91d75" exitCode=0 Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.176041 4809 generic.go:334] "Generic (PLEG): container finished" podID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerID="bd8eec56078ff1a5910c47c787dadb501a8f077b8412160d3774e70541b98b24" exitCode=2 Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.176055 4809 generic.go:334] "Generic (PLEG): container finished" podID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerID="c52a6040f3804de9369eccd5891abcf98bff5e7d1b87d9f9a1ceb591946cad99" exitCode=0 Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.175720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerDied","Data":"4642b43a93b98e574260a834a13c519c1eda1ddb903a08eccc2d908cada91d75"} Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.176098 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerDied","Data":"bd8eec56078ff1a5910c47c787dadb501a8f077b8412160d3774e70541b98b24"} Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.176115 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerDied","Data":"c52a6040f3804de9369eccd5891abcf98bff5e7d1b87d9f9a1ceb591946cad99"} Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.178397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3ff647e6-3fcb-401c-a695-047f07ca9a58","Type":"ContainerStarted","Data":"94d8942a8db6962cfb51843e4164ce1943cc0c425f97a5f3472675b030ed50f8"} Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.178418 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3ff647e6-3fcb-401c-a695-047f07ca9a58","Type":"ContainerStarted","Data":"da1b3f07b5cff1c4f80c624f92dcdee402da5120b1d3bfd01acf618e9d081f5b"} Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.178559 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.197779 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.8398562680000001 podStartE2EDuration="2.197733438s" podCreationTimestamp="2025-11-27 17:41:09 +0000 UTC" firstStartedPulling="2025-11-27 17:41:10.203520542 +0000 UTC m=+1905.475977894" lastFinishedPulling="2025-11-27 17:41:10.561397712 +0000 UTC m=+1905.833855064" observedRunningTime="2025-11-27 17:41:11.195146107 +0000 UTC m=+1906.467603469" watchObservedRunningTime="2025-11-27 17:41:11.197733438 +0000 UTC m=+1906.470190800" Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.731983 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 27 17:41:11 crc kubenswrapper[4809]: I1127 17:41:11.764652 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 27 17:41:12 crc kubenswrapper[4809]: I1127 17:41:12.227489 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 27 17:41:12 crc kubenswrapper[4809]: I1127 17:41:12.490782 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 17:41:12 crc kubenswrapper[4809]: I1127 17:41:12.492213 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 17:41:13 crc kubenswrapper[4809]: I1127 17:41:13.574950 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.209:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 17:41:13 crc kubenswrapper[4809]: I1127 17:41:13.574950 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.209:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.211898 4809 generic.go:334] "Generic (PLEG): container finished" podID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerID="162fdb74efc4fa762b0e1378da20e568bcd84b3daeea002661753778e9f08150" exitCode=0 Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.212097 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerDied","Data":"162fdb74efc4fa762b0e1378da20e568bcd84b3daeea002661753778e9f08150"} Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.541265 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.579609 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-config-data\") pod \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.580062 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-log-httpd\") pod \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.580140 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-combined-ca-bundle\") pod \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.580313 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-sg-core-conf-yaml\") pod \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.580343 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6mht\" (UniqueName: \"kubernetes.io/projected/f802a83a-cdca-4a71-911c-63dc4c8f14f6-kube-api-access-r6mht\") pod \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.580405 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-run-httpd\") pod \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.580456 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-scripts\") pod \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\" (UID: \"f802a83a-cdca-4a71-911c-63dc4c8f14f6\") " Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.583375 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f802a83a-cdca-4a71-911c-63dc4c8f14f6" (UID: "f802a83a-cdca-4a71-911c-63dc4c8f14f6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.586608 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f802a83a-cdca-4a71-911c-63dc4c8f14f6-kube-api-access-r6mht" (OuterVolumeSpecName: "kube-api-access-r6mht") pod "f802a83a-cdca-4a71-911c-63dc4c8f14f6" (UID: "f802a83a-cdca-4a71-911c-63dc4c8f14f6"). InnerVolumeSpecName "kube-api-access-r6mht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.587036 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f802a83a-cdca-4a71-911c-63dc4c8f14f6" (UID: "f802a83a-cdca-4a71-911c-63dc4c8f14f6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.587175 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-scripts" (OuterVolumeSpecName: "scripts") pod "f802a83a-cdca-4a71-911c-63dc4c8f14f6" (UID: "f802a83a-cdca-4a71-911c-63dc4c8f14f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.631936 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f802a83a-cdca-4a71-911c-63dc4c8f14f6" (UID: "f802a83a-cdca-4a71-911c-63dc4c8f14f6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.682958 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.682993 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6mht\" (UniqueName: \"kubernetes.io/projected/f802a83a-cdca-4a71-911c-63dc4c8f14f6-kube-api-access-r6mht\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.683011 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.683024 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.683036 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f802a83a-cdca-4a71-911c-63dc4c8f14f6-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.684064 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f802a83a-cdca-4a71-911c-63dc4c8f14f6" (UID: "f802a83a-cdca-4a71-911c-63dc4c8f14f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.720801 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-config-data" (OuterVolumeSpecName: "config-data") pod "f802a83a-cdca-4a71-911c-63dc4c8f14f6" (UID: "f802a83a-cdca-4a71-911c-63dc4c8f14f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.784528 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:14 crc kubenswrapper[4809]: I1127 17:41:14.784565 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f802a83a-cdca-4a71-911c-63dc4c8f14f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.228461 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f802a83a-cdca-4a71-911c-63dc4c8f14f6","Type":"ContainerDied","Data":"f7c7348e8ada464abee6949902a178e7a365dcbac8e78c4068f6ac3767000b8f"} Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.228519 4809 scope.go:117] "RemoveContainer" containerID="4642b43a93b98e574260a834a13c519c1eda1ddb903a08eccc2d908cada91d75" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.228713 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.263908 4809 scope.go:117] "RemoveContainer" containerID="bd8eec56078ff1a5910c47c787dadb501a8f077b8412160d3774e70541b98b24" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.270283 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.283431 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.304538 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:15 crc kubenswrapper[4809]: E1127 17:41:15.304971 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="sg-core" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.304988 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="sg-core" Nov 27 17:41:15 crc kubenswrapper[4809]: E1127 17:41:15.305011 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="ceilometer-notification-agent" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.305017 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="ceilometer-notification-agent" Nov 27 17:41:15 crc kubenswrapper[4809]: E1127 17:41:15.305030 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="proxy-httpd" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.305037 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="proxy-httpd" Nov 27 17:41:15 crc kubenswrapper[4809]: E1127 17:41:15.305046 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="ceilometer-central-agent" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.305052 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="ceilometer-central-agent" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.305228 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="ceilometer-central-agent" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.305243 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="sg-core" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.305260 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="proxy-httpd" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.305270 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" containerName="ceilometer-notification-agent" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.312470 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: E1127 17:41:15.316389 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf802a83a_cdca_4a71_911c_63dc4c8f14f6.slice\": RecentStats: unable to find data in memory cache]" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.326706 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.326960 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.328475 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.339809 4809 scope.go:117] "RemoveContainer" containerID="162fdb74efc4fa762b0e1378da20e568bcd84b3daeea002661753778e9f08150" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.357843 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.370112 4809 scope.go:117] "RemoveContainer" containerID="c52a6040f3804de9369eccd5891abcf98bff5e7d1b87d9f9a1ceb591946cad99" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.394255 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-log-httpd\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.394342 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.394377 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-config-data\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.394396 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndr4x\" (UniqueName: \"kubernetes.io/projected/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-kube-api-access-ndr4x\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.394413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.394451 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-run-httpd\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.394473 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.394500 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-scripts\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.469382 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f802a83a-cdca-4a71-911c-63dc4c8f14f6" path="/var/lib/kubelet/pods/f802a83a-cdca-4a71-911c-63dc4c8f14f6/volumes" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.496706 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.496803 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-config-data\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.496829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndr4x\" (UniqueName: \"kubernetes.io/projected/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-kube-api-access-ndr4x\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.496855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.496938 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-run-httpd\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.496962 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.496991 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-scripts\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.497069 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-log-httpd\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.498006 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-log-httpd\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.498105 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-run-httpd\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.501582 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.501709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.504806 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-scripts\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.511809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-config-data\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.513454 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.519899 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndr4x\" (UniqueName: \"kubernetes.io/projected/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-kube-api-access-ndr4x\") pod \"ceilometer-0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " pod="openstack/ceilometer-0" Nov 27 17:41:15 crc kubenswrapper[4809]: I1127 17:41:15.633829 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:16 crc kubenswrapper[4809]: W1127 17:41:16.120654 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27d1e1a2_b449_44e2_a6f0_c2543a8bd9e0.slice/crio-34bda3e293bd3710a09f2c03337bba8692085c3c244729586b64d8e894cf5b01 WatchSource:0}: Error finding container 34bda3e293bd3710a09f2c03337bba8692085c3c244729586b64d8e894cf5b01: Status 404 returned error can't find the container with id 34bda3e293bd3710a09f2c03337bba8692085c3c244729586b64d8e894cf5b01 Nov 27 17:41:16 crc kubenswrapper[4809]: I1127 17:41:16.124362 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:16 crc kubenswrapper[4809]: I1127 17:41:16.239731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerStarted","Data":"34bda3e293bd3710a09f2c03337bba8692085c3c244729586b64d8e894cf5b01"} Nov 27 17:41:17 crc kubenswrapper[4809]: I1127 17:41:17.255231 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerStarted","Data":"90db778bcf044c1785f6e1978c00c6990a20fb904e30cf0f1825f0a35a91bce4"} Nov 27 17:41:18 crc kubenswrapper[4809]: I1127 17:41:18.267569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerStarted","Data":"f335bdff234e4de366d4c4edd0bfdd260088a80ca30373dea2082565763caf63"} Nov 27 17:41:18 crc kubenswrapper[4809]: I1127 17:41:18.429327 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 17:41:18 crc kubenswrapper[4809]: I1127 17:41:18.429949 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 17:41:18 crc kubenswrapper[4809]: I1127 17:41:18.436623 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 17:41:19 crc kubenswrapper[4809]: I1127 17:41:19.279557 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerStarted","Data":"f43673b1c845ebfb6920e34289c3c2edff52e86ddecccecc6c68c104224c98e6"} Nov 27 17:41:19 crc kubenswrapper[4809]: I1127 17:41:19.284267 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 17:41:19 crc kubenswrapper[4809]: I1127 17:41:19.583963 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.293221 4809 generic.go:334] "Generic (PLEG): container finished" podID="0ffdcceb-d220-4a44-88f9-94f7caf3694f" containerID="e0d64cac60b1d7dd70920d451042a26871b366b24598244ca13f72d1dea7dc9b" exitCode=137 Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.293361 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0ffdcceb-d220-4a44-88f9-94f7caf3694f","Type":"ContainerDied","Data":"e0d64cac60b1d7dd70920d451042a26871b366b24598244ca13f72d1dea7dc9b"} Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.293643 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0ffdcceb-d220-4a44-88f9-94f7caf3694f","Type":"ContainerDied","Data":"b08842eeeb2dd204f7b9564d27fc89861a2b60024c147ee0e8be2997c55f4ad1"} Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.293662 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b08842eeeb2dd204f7b9564d27fc89861a2b60024c147ee0e8be2997c55f4ad1" Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.340705 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.391291 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-combined-ca-bundle\") pod \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.391375 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n7rk\" (UniqueName: \"kubernetes.io/projected/0ffdcceb-d220-4a44-88f9-94f7caf3694f-kube-api-access-6n7rk\") pod \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.391545 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-config-data\") pod \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\" (UID: \"0ffdcceb-d220-4a44-88f9-94f7caf3694f\") " Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.399107 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ffdcceb-d220-4a44-88f9-94f7caf3694f-kube-api-access-6n7rk" (OuterVolumeSpecName: "kube-api-access-6n7rk") pod "0ffdcceb-d220-4a44-88f9-94f7caf3694f" (UID: "0ffdcceb-d220-4a44-88f9-94f7caf3694f"). InnerVolumeSpecName "kube-api-access-6n7rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.424926 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ffdcceb-d220-4a44-88f9-94f7caf3694f" (UID: "0ffdcceb-d220-4a44-88f9-94f7caf3694f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.437399 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-config-data" (OuterVolumeSpecName: "config-data") pod "0ffdcceb-d220-4a44-88f9-94f7caf3694f" (UID: "0ffdcceb-d220-4a44-88f9-94f7caf3694f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.495020 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.495298 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffdcceb-d220-4a44-88f9-94f7caf3694f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:20 crc kubenswrapper[4809]: I1127 17:41:20.495317 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n7rk\" (UniqueName: \"kubernetes.io/projected/0ffdcceb-d220-4a44-88f9-94f7caf3694f-kube-api-access-6n7rk\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.310450 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.318832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerStarted","Data":"4c6eb6b65559e191b18a64bea72686a2ce1ea19000a4e9fd31c2065922bceb93"} Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.318914 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.340532 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.322645257 podStartE2EDuration="6.340507432s" podCreationTimestamp="2025-11-27 17:41:15 +0000 UTC" firstStartedPulling="2025-11-27 17:41:16.123084047 +0000 UTC m=+1911.395541399" lastFinishedPulling="2025-11-27 17:41:20.140946222 +0000 UTC m=+1915.413403574" observedRunningTime="2025-11-27 17:41:21.332771953 +0000 UTC m=+1916.605229315" watchObservedRunningTime="2025-11-27 17:41:21.340507432 +0000 UTC m=+1916.612964784" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.357008 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.371824 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.384233 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 17:41:21 crc kubenswrapper[4809]: E1127 17:41:21.384819 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffdcceb-d220-4a44-88f9-94f7caf3694f" containerName="nova-cell1-novncproxy-novncproxy" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.384844 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffdcceb-d220-4a44-88f9-94f7caf3694f" containerName="nova-cell1-novncproxy-novncproxy" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.385139 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ffdcceb-d220-4a44-88f9-94f7caf3694f" containerName="nova-cell1-novncproxy-novncproxy" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.386469 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.389925 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.390150 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.390318 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.423226 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.423317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.423424 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.426708 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqppk\" (UniqueName: \"kubernetes.io/projected/969b1f1d-ce0b-48fd-b149-72649f406cf0-kube-api-access-nqppk\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.426969 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.434507 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.483243 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ffdcceb-d220-4a44-88f9-94f7caf3694f" path="/var/lib/kubelet/pods/0ffdcceb-d220-4a44-88f9-94f7caf3694f/volumes" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.528200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqppk\" (UniqueName: \"kubernetes.io/projected/969b1f1d-ce0b-48fd-b149-72649f406cf0-kube-api-access-nqppk\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.528267 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.528348 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.528375 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.528415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.532406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.532797 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.533234 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.543420 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/969b1f1d-ce0b-48fd-b149-72649f406cf0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.544115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqppk\" (UniqueName: \"kubernetes.io/projected/969b1f1d-ce0b-48fd-b149-72649f406cf0-kube-api-access-nqppk\") pod \"nova-cell1-novncproxy-0\" (UID: \"969b1f1d-ce0b-48fd-b149-72649f406cf0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:21 crc kubenswrapper[4809]: I1127 17:41:21.727567 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:22 crc kubenswrapper[4809]: I1127 17:41:22.189696 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 17:41:22 crc kubenswrapper[4809]: W1127 17:41:22.195200 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod969b1f1d_ce0b_48fd_b149_72649f406cf0.slice/crio-37ce636c22198feb54ee5439865841657c28041b18aa182424c83ee876028693 WatchSource:0}: Error finding container 37ce636c22198feb54ee5439865841657c28041b18aa182424c83ee876028693: Status 404 returned error can't find the container with id 37ce636c22198feb54ee5439865841657c28041b18aa182424c83ee876028693 Nov 27 17:41:22 crc kubenswrapper[4809]: I1127 17:41:22.321698 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"969b1f1d-ce0b-48fd-b149-72649f406cf0","Type":"ContainerStarted","Data":"37ce636c22198feb54ee5439865841657c28041b18aa182424c83ee876028693"} Nov 27 17:41:22 crc kubenswrapper[4809]: I1127 17:41:22.494345 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 17:41:22 crc kubenswrapper[4809]: I1127 17:41:22.494812 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 17:41:22 crc kubenswrapper[4809]: I1127 17:41:22.495682 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 17:41:22 crc kubenswrapper[4809]: I1127 17:41:22.500273 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.333517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"969b1f1d-ce0b-48fd-b149-72649f406cf0","Type":"ContainerStarted","Data":"fe544a769f61b9ae968ed0ebc532176161930784dc253f53250511cd4cde9bfa"} Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.333562 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.337863 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.362932 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.362913835 podStartE2EDuration="2.362913835s" podCreationTimestamp="2025-11-27 17:41:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:23.355230926 +0000 UTC m=+1918.627688288" watchObservedRunningTime="2025-11-27 17:41:23.362913835 +0000 UTC m=+1918.635371177" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.468051 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:41:23 crc kubenswrapper[4809]: E1127 17:41:23.468298 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.525300 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-tcmnm"] Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.528477 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.564504 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-tcmnm"] Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.570662 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-config\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.570845 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.570892 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.570916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.570935 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgpl6\" (UniqueName: \"kubernetes.io/projected/14a8fdea-bcf5-49a2-8b88-13875ac05c71-kube-api-access-sgpl6\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.570972 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.673070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.673130 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.673158 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.673180 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgpl6\" (UniqueName: \"kubernetes.io/projected/14a8fdea-bcf5-49a2-8b88-13875ac05c71-kube-api-access-sgpl6\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.673207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.673278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-config\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.674015 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.674169 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.674214 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-config\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.674486 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.675116 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.691560 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgpl6\" (UniqueName: \"kubernetes.io/projected/14a8fdea-bcf5-49a2-8b88-13875ac05c71-kube-api-access-sgpl6\") pod \"dnsmasq-dns-59cf4bdb65-tcmnm\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:23 crc kubenswrapper[4809]: I1127 17:41:23.849660 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:24 crc kubenswrapper[4809]: I1127 17:41:24.478980 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-tcmnm"] Nov 27 17:41:25 crc kubenswrapper[4809]: I1127 17:41:25.362062 4809 generic.go:334] "Generic (PLEG): container finished" podID="14a8fdea-bcf5-49a2-8b88-13875ac05c71" containerID="d534bd6bf4808e319567a79fb1432f5d93ad40467433da850231a465b186b97a" exitCode=0 Nov 27 17:41:25 crc kubenswrapper[4809]: I1127 17:41:25.362150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" event={"ID":"14a8fdea-bcf5-49a2-8b88-13875ac05c71","Type":"ContainerDied","Data":"d534bd6bf4808e319567a79fb1432f5d93ad40467433da850231a465b186b97a"} Nov 27 17:41:25 crc kubenswrapper[4809]: I1127 17:41:25.362480 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" event={"ID":"14a8fdea-bcf5-49a2-8b88-13875ac05c71","Type":"ContainerStarted","Data":"953256730a2a0506a772c41863723363afc6f72239d4ce5f07bf7b4c2ab45f50"} Nov 27 17:41:25 crc kubenswrapper[4809]: I1127 17:41:25.861581 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:25 crc kubenswrapper[4809]: I1127 17:41:25.862471 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="proxy-httpd" containerID="cri-o://4c6eb6b65559e191b18a64bea72686a2ce1ea19000a4e9fd31c2065922bceb93" gracePeriod=30 Nov 27 17:41:25 crc kubenswrapper[4809]: I1127 17:41:25.862533 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="sg-core" containerID="cri-o://f43673b1c845ebfb6920e34289c3c2edff52e86ddecccecc6c68c104224c98e6" gracePeriod=30 Nov 27 17:41:25 crc kubenswrapper[4809]: I1127 17:41:25.862436 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="ceilometer-central-agent" containerID="cri-o://90db778bcf044c1785f6e1978c00c6990a20fb904e30cf0f1825f0a35a91bce4" gracePeriod=30 Nov 27 17:41:25 crc kubenswrapper[4809]: I1127 17:41:25.862570 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="ceilometer-notification-agent" containerID="cri-o://f335bdff234e4de366d4c4edd0bfdd260088a80ca30373dea2082565763caf63" gracePeriod=30 Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.389056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" event={"ID":"14a8fdea-bcf5-49a2-8b88-13875ac05c71","Type":"ContainerStarted","Data":"487c143437edb8bde10d7fd825cbecf6c0f79773064bb64afa371599825ed8ac"} Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.389438 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.394050 4809 generic.go:334] "Generic (PLEG): container finished" podID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerID="4c6eb6b65559e191b18a64bea72686a2ce1ea19000a4e9fd31c2065922bceb93" exitCode=0 Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.394260 4809 generic.go:334] "Generic (PLEG): container finished" podID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerID="f43673b1c845ebfb6920e34289c3c2edff52e86ddecccecc6c68c104224c98e6" exitCode=2 Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.394272 4809 generic.go:334] "Generic (PLEG): container finished" podID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerID="f335bdff234e4de366d4c4edd0bfdd260088a80ca30373dea2082565763caf63" exitCode=0 Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.394281 4809 generic.go:334] "Generic (PLEG): container finished" podID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerID="90db778bcf044c1785f6e1978c00c6990a20fb904e30cf0f1825f0a35a91bce4" exitCode=0 Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.394303 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerDied","Data":"4c6eb6b65559e191b18a64bea72686a2ce1ea19000a4e9fd31c2065922bceb93"} Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.394518 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerDied","Data":"f43673b1c845ebfb6920e34289c3c2edff52e86ddecccecc6c68c104224c98e6"} Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.394536 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerDied","Data":"f335bdff234e4de366d4c4edd0bfdd260088a80ca30373dea2082565763caf63"} Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.394548 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerDied","Data":"90db778bcf044c1785f6e1978c00c6990a20fb904e30cf0f1825f0a35a91bce4"} Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.413674 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" podStartSLOduration=3.413654877 podStartE2EDuration="3.413654877s" podCreationTimestamp="2025-11-27 17:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:26.410470121 +0000 UTC m=+1921.682927483" watchObservedRunningTime="2025-11-27 17:41:26.413654877 +0000 UTC m=+1921.686112229" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.706579 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.727723 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846004 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-run-httpd\") pod \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846041 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-scripts\") pod \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846108 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-combined-ca-bundle\") pod \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846137 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-log-httpd\") pod \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846157 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-ceilometer-tls-certs\") pod \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846250 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-config-data\") pod \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846277 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndr4x\" (UniqueName: \"kubernetes.io/projected/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-kube-api-access-ndr4x\") pod \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846325 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" (UID: "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846367 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-sg-core-conf-yaml\") pod \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\" (UID: \"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0\") " Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846572 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" (UID: "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846904 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.846931 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.851526 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-scripts" (OuterVolumeSpecName: "scripts") pod "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" (UID: "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.854032 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-kube-api-access-ndr4x" (OuterVolumeSpecName: "kube-api-access-ndr4x") pod "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" (UID: "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0"). InnerVolumeSpecName "kube-api-access-ndr4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.888197 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" (UID: "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.912536 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" (UID: "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.931410 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" (UID: "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.947967 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndr4x\" (UniqueName: \"kubernetes.io/projected/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-kube-api-access-ndr4x\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.947999 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.948009 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.948018 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.948026 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.977530 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-config-data" (OuterVolumeSpecName: "config-data") pod "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" (UID: "27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.991406 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.991663 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-log" containerID="cri-o://49dcdb1d225f90d0f63215dd008123276303a47ad31fafddf95d0ab07dffbcdd" gracePeriod=30 Nov 27 17:41:26 crc kubenswrapper[4809]: I1127 17:41:26.991776 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-api" containerID="cri-o://5cecd8bc240027b6b665d7bd337bc08fa70cf6bdf17404efc8a9e9b7ad749ba2" gracePeriod=30 Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.049250 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.409506 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0","Type":"ContainerDied","Data":"34bda3e293bd3710a09f2c03337bba8692085c3c244729586b64d8e894cf5b01"} Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.409558 4809 scope.go:117] "RemoveContainer" containerID="4c6eb6b65559e191b18a64bea72686a2ce1ea19000a4e9fd31c2065922bceb93" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.409687 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.416809 4809 generic.go:334] "Generic (PLEG): container finished" podID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerID="49dcdb1d225f90d0f63215dd008123276303a47ad31fafddf95d0ab07dffbcdd" exitCode=143 Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.417408 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1895833b-2f96-438e-8ab2-85b1d9752b87","Type":"ContainerDied","Data":"49dcdb1d225f90d0f63215dd008123276303a47ad31fafddf95d0ab07dffbcdd"} Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.432318 4809 scope.go:117] "RemoveContainer" containerID="f43673b1c845ebfb6920e34289c3c2edff52e86ddecccecc6c68c104224c98e6" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.447347 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.461987 4809 scope.go:117] "RemoveContainer" containerID="f335bdff234e4de366d4c4edd0bfdd260088a80ca30373dea2082565763caf63" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.483868 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.483910 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:27 crc kubenswrapper[4809]: E1127 17:41:27.484980 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="sg-core" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.485002 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="sg-core" Nov 27 17:41:27 crc kubenswrapper[4809]: E1127 17:41:27.485026 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="ceilometer-central-agent" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.485033 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="ceilometer-central-agent" Nov 27 17:41:27 crc kubenswrapper[4809]: E1127 17:41:27.485040 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="ceilometer-notification-agent" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.485046 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="ceilometer-notification-agent" Nov 27 17:41:27 crc kubenswrapper[4809]: E1127 17:41:27.485065 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="proxy-httpd" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.485070 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="proxy-httpd" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.485263 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="proxy-httpd" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.485287 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="ceilometer-notification-agent" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.485302 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="ceilometer-central-agent" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.485314 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" containerName="sg-core" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.487167 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.488575 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.491581 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.491694 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.493197 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.517694 4809 scope.go:117] "RemoveContainer" containerID="90db778bcf044c1785f6e1978c00c6990a20fb904e30cf0f1825f0a35a91bce4" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.663415 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-log-httpd\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.663468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpp8w\" (UniqueName: \"kubernetes.io/projected/8ae721b0-5274-4f53-a9de-4942464bd7a5-kube-api-access-hpp8w\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.663580 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-scripts\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.663609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-run-httpd\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.663706 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.663775 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-config-data\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.663801 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.663833 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765068 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-log-httpd\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765117 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpp8w\" (UniqueName: \"kubernetes.io/projected/8ae721b0-5274-4f53-a9de-4942464bd7a5-kube-api-access-hpp8w\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-scripts\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-run-httpd\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765242 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765268 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-config-data\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765288 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765309 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765564 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-log-httpd\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.765865 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-run-httpd\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.769902 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.769950 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.771045 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-config-data\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.773549 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-scripts\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.787071 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpp8w\" (UniqueName: \"kubernetes.io/projected/8ae721b0-5274-4f53-a9de-4942464bd7a5-kube-api-access-hpp8w\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.788319 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " pod="openstack/ceilometer-0" Nov 27 17:41:27 crc kubenswrapper[4809]: I1127 17:41:27.816278 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:28 crc kubenswrapper[4809]: I1127 17:41:28.934151 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:28 crc kubenswrapper[4809]: W1127 17:41:28.935932 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ae721b0_5274_4f53_a9de_4942464bd7a5.slice/crio-4b7394d563dd5bc7a31ee8cc71ebe0915c02e55bd0da75155c29b8892676881c WatchSource:0}: Error finding container 4b7394d563dd5bc7a31ee8cc71ebe0915c02e55bd0da75155c29b8892676881c: Status 404 returned error can't find the container with id 4b7394d563dd5bc7a31ee8cc71ebe0915c02e55bd0da75155c29b8892676881c Nov 27 17:41:29 crc kubenswrapper[4809]: I1127 17:41:29.262792 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:29 crc kubenswrapper[4809]: I1127 17:41:29.443930 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerStarted","Data":"4b7394d563dd5bc7a31ee8cc71ebe0915c02e55bd0da75155c29b8892676881c"} Nov 27 17:41:29 crc kubenswrapper[4809]: I1127 17:41:29.470230 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0" path="/var/lib/kubelet/pods/27d1e1a2-b449-44e2-a6f0-c2543a8bd9e0/volumes" Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.460957 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerStarted","Data":"f941ed7cd0538dfe1787e6d20a8f527b30cc4c43cd05b6efe44e48e81d763d61"} Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.468880 4809 generic.go:334] "Generic (PLEG): container finished" podID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerID="5cecd8bc240027b6b665d7bd337bc08fa70cf6bdf17404efc8a9e9b7ad749ba2" exitCode=0 Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.468918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1895833b-2f96-438e-8ab2-85b1d9752b87","Type":"ContainerDied","Data":"5cecd8bc240027b6b665d7bd337bc08fa70cf6bdf17404efc8a9e9b7ad749ba2"} Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.693537 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.824199 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1895833b-2f96-438e-8ab2-85b1d9752b87-logs\") pod \"1895833b-2f96-438e-8ab2-85b1d9752b87\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.824376 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp2z9\" (UniqueName: \"kubernetes.io/projected/1895833b-2f96-438e-8ab2-85b1d9752b87-kube-api-access-mp2z9\") pod \"1895833b-2f96-438e-8ab2-85b1d9752b87\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.824416 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-combined-ca-bundle\") pod \"1895833b-2f96-438e-8ab2-85b1d9752b87\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.824562 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-config-data\") pod \"1895833b-2f96-438e-8ab2-85b1d9752b87\" (UID: \"1895833b-2f96-438e-8ab2-85b1d9752b87\") " Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.824879 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1895833b-2f96-438e-8ab2-85b1d9752b87-logs" (OuterVolumeSpecName: "logs") pod "1895833b-2f96-438e-8ab2-85b1d9752b87" (UID: "1895833b-2f96-438e-8ab2-85b1d9752b87"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.825813 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1895833b-2f96-438e-8ab2-85b1d9752b87-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.842257 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1895833b-2f96-438e-8ab2-85b1d9752b87-kube-api-access-mp2z9" (OuterVolumeSpecName: "kube-api-access-mp2z9") pod "1895833b-2f96-438e-8ab2-85b1d9752b87" (UID: "1895833b-2f96-438e-8ab2-85b1d9752b87"). InnerVolumeSpecName "kube-api-access-mp2z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.889057 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-config-data" (OuterVolumeSpecName: "config-data") pod "1895833b-2f96-438e-8ab2-85b1d9752b87" (UID: "1895833b-2f96-438e-8ab2-85b1d9752b87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.891054 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1895833b-2f96-438e-8ab2-85b1d9752b87" (UID: "1895833b-2f96-438e-8ab2-85b1d9752b87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.928696 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp2z9\" (UniqueName: \"kubernetes.io/projected/1895833b-2f96-438e-8ab2-85b1d9752b87-kube-api-access-mp2z9\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.928765 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:30 crc kubenswrapper[4809]: I1127 17:41:30.928782 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1895833b-2f96-438e-8ab2-85b1d9752b87-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.480604 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerStarted","Data":"8b49173719314457e7899b6f9cf63a1bd5e7c69dac6c00349c1d4d7788eb3d44"} Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.480665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerStarted","Data":"73f9da341bfe10da04849f02b3fa14b390b234203c7af032dc113e2521b6cb8a"} Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.484115 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1895833b-2f96-438e-8ab2-85b1d9752b87","Type":"ContainerDied","Data":"2998d53f72aabdcba55037edf00afabf5505720f2485719642a0f68648927a20"} Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.484169 4809 scope.go:117] "RemoveContainer" containerID="5cecd8bc240027b6b665d7bd337bc08fa70cf6bdf17404efc8a9e9b7ad749ba2" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.484175 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.522869 4809 scope.go:117] "RemoveContainer" containerID="49dcdb1d225f90d0f63215dd008123276303a47ad31fafddf95d0ab07dffbcdd" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.526170 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.535223 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.556215 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:31 crc kubenswrapper[4809]: E1127 17:41:31.556588 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-log" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.556604 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-log" Nov 27 17:41:31 crc kubenswrapper[4809]: E1127 17:41:31.556629 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-api" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.556635 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-api" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.556923 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-api" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.556957 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" containerName="nova-api-log" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.559714 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.562071 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.562352 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.562461 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.574833 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.728139 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.745942 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-config-data\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.746176 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-public-tls-certs\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.746250 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/887efa58-415d-4bcc-b818-09693edd49f3-logs\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.746302 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr6mc\" (UniqueName: \"kubernetes.io/projected/887efa58-415d-4bcc-b818-09693edd49f3-kube-api-access-qr6mc\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.746348 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.746445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.759149 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.848124 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/887efa58-415d-4bcc-b818-09693edd49f3-logs\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.848186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr6mc\" (UniqueName: \"kubernetes.io/projected/887efa58-415d-4bcc-b818-09693edd49f3-kube-api-access-qr6mc\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.848217 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.848265 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.848345 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-config-data\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.848505 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-public-tls-certs\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.850031 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/887efa58-415d-4bcc-b818-09693edd49f3-logs\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.854965 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.854988 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-config-data\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.857905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.863282 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-public-tls-certs\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.868431 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr6mc\" (UniqueName: \"kubernetes.io/projected/887efa58-415d-4bcc-b818-09693edd49f3-kube-api-access-qr6mc\") pod \"nova-api-0\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " pod="openstack/nova-api-0" Nov 27 17:41:31 crc kubenswrapper[4809]: I1127 17:41:31.879637 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.346304 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:32 crc kubenswrapper[4809]: W1127 17:41:32.348837 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod887efa58_415d_4bcc_b818_09693edd49f3.slice/crio-208f30a77604ed58b3a65b5262318cd40db82dae6bec058081f9df939c9f1643 WatchSource:0}: Error finding container 208f30a77604ed58b3a65b5262318cd40db82dae6bec058081f9df939c9f1643: Status 404 returned error can't find the container with id 208f30a77604ed58b3a65b5262318cd40db82dae6bec058081f9df939c9f1643 Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.496252 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"887efa58-415d-4bcc-b818-09693edd49f3","Type":"ContainerStarted","Data":"208f30a77604ed58b3a65b5262318cd40db82dae6bec058081f9df939c9f1643"} Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.518682 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.667396 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-rbmz5"] Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.669182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.677206 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.677394 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.697065 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rbmz5"] Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.770508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgxv9\" (UniqueName: \"kubernetes.io/projected/79ca60f5-37e1-4870-8282-7bcb50174a9e-kube-api-access-pgxv9\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.770649 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-scripts\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.770688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-config-data\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.770721 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.872435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-scripts\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.872531 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-config-data\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.872564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.872676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgxv9\" (UniqueName: \"kubernetes.io/projected/79ca60f5-37e1-4870-8282-7bcb50174a9e-kube-api-access-pgxv9\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.877886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.878244 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-scripts\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.883054 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-config-data\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:32 crc kubenswrapper[4809]: I1127 17:41:32.896293 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgxv9\" (UniqueName: \"kubernetes.io/projected/79ca60f5-37e1-4870-8282-7bcb50174a9e-kube-api-access-pgxv9\") pod \"nova-cell1-cell-mapping-rbmz5\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:33 crc kubenswrapper[4809]: I1127 17:41:33.132211 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:33 crc kubenswrapper[4809]: I1127 17:41:33.481222 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1895833b-2f96-438e-8ab2-85b1d9752b87" path="/var/lib/kubelet/pods/1895833b-2f96-438e-8ab2-85b1d9752b87/volumes" Nov 27 17:41:33 crc kubenswrapper[4809]: I1127 17:41:33.506945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"887efa58-415d-4bcc-b818-09693edd49f3","Type":"ContainerStarted","Data":"38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13"} Nov 27 17:41:33 crc kubenswrapper[4809]: I1127 17:41:33.507031 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"887efa58-415d-4bcc-b818-09693edd49f3","Type":"ContainerStarted","Data":"9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826"} Nov 27 17:41:33 crc kubenswrapper[4809]: I1127 17:41:33.528385 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.528362444 podStartE2EDuration="2.528362444s" podCreationTimestamp="2025-11-27 17:41:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:33.526118153 +0000 UTC m=+1928.798575495" watchObservedRunningTime="2025-11-27 17:41:33.528362444 +0000 UTC m=+1928.800819816" Nov 27 17:41:33 crc kubenswrapper[4809]: I1127 17:41:33.659877 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rbmz5"] Nov 27 17:41:33 crc kubenswrapper[4809]: W1127 17:41:33.661393 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79ca60f5_37e1_4870_8282_7bcb50174a9e.slice/crio-2434fe16280fe5dea9b343f3d20025a110924f688af6cfd8357f05b2ba9498f0 WatchSource:0}: Error finding container 2434fe16280fe5dea9b343f3d20025a110924f688af6cfd8357f05b2ba9498f0: Status 404 returned error can't find the container with id 2434fe16280fe5dea9b343f3d20025a110924f688af6cfd8357f05b2ba9498f0 Nov 27 17:41:33 crc kubenswrapper[4809]: I1127 17:41:33.851632 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:41:33 crc kubenswrapper[4809]: I1127 17:41:33.929159 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xtlg4"] Nov 27 17:41:33 crc kubenswrapper[4809]: I1127 17:41:33.929459 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" podUID="16fbb102-acdd-49fd-a941-9e7ff32b71cc" containerName="dnsmasq-dns" containerID="cri-o://84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490" gracePeriod=10 Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.460641 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:41:34 crc kubenswrapper[4809]: E1127 17:41:34.461366 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.495035 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.522276 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rbmz5" event={"ID":"79ca60f5-37e1-4870-8282-7bcb50174a9e","Type":"ContainerStarted","Data":"3df4fcce6640f111dad80a2d5cb0be2f02cd5c4ad1125a7dd6e4e1105e60decc"} Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.522319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rbmz5" event={"ID":"79ca60f5-37e1-4870-8282-7bcb50174a9e","Type":"ContainerStarted","Data":"2434fe16280fe5dea9b343f3d20025a110924f688af6cfd8357f05b2ba9498f0"} Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.528359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerStarted","Data":"3b10e9f2ef60c91e673f65348b8a867d006b7716d7933446f09153b94268b578"} Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.528694 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="ceilometer-central-agent" containerID="cri-o://f941ed7cd0538dfe1787e6d20a8f527b30cc4c43cd05b6efe44e48e81d763d61" gracePeriod=30 Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.528826 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.528891 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="ceilometer-notification-agent" containerID="cri-o://73f9da341bfe10da04849f02b3fa14b390b234203c7af032dc113e2521b6cb8a" gracePeriod=30 Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.528910 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="proxy-httpd" containerID="cri-o://3b10e9f2ef60c91e673f65348b8a867d006b7716d7933446f09153b94268b578" gracePeriod=30 Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.528902 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="sg-core" containerID="cri-o://8b49173719314457e7899b6f9cf63a1bd5e7c69dac6c00349c1d4d7788eb3d44" gracePeriod=30 Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.535516 4809 generic.go:334] "Generic (PLEG): container finished" podID="16fbb102-acdd-49fd-a941-9e7ff32b71cc" containerID="84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490" exitCode=0 Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.535907 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" event={"ID":"16fbb102-acdd-49fd-a941-9e7ff32b71cc","Type":"ContainerDied","Data":"84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490"} Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.535944 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" event={"ID":"16fbb102-acdd-49fd-a941-9e7ff32b71cc","Type":"ContainerDied","Data":"9b4e73bb51e470d478112e8c3162fb03ad704e09a83cc7085c954d12afba6832"} Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.535965 4809 scope.go:117] "RemoveContainer" containerID="84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.544257 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xtlg4" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.552896 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-rbmz5" podStartSLOduration=2.55287711 podStartE2EDuration="2.55287711s" podCreationTimestamp="2025-11-27 17:41:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:34.540516626 +0000 UTC m=+1929.812973988" watchObservedRunningTime="2025-11-27 17:41:34.55287711 +0000 UTC m=+1929.825334462" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.567236 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.050111954 podStartE2EDuration="7.567217019s" podCreationTimestamp="2025-11-27 17:41:27 +0000 UTC" firstStartedPulling="2025-11-27 17:41:28.939019731 +0000 UTC m=+1924.211477083" lastFinishedPulling="2025-11-27 17:41:33.456124796 +0000 UTC m=+1928.728582148" observedRunningTime="2025-11-27 17:41:34.560830626 +0000 UTC m=+1929.833287988" watchObservedRunningTime="2025-11-27 17:41:34.567217019 +0000 UTC m=+1929.839674371" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.577022 4809 scope.go:117] "RemoveContainer" containerID="d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.612943 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-sb\") pod \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.613072 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-nb\") pod \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.613104 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27cjl\" (UniqueName: \"kubernetes.io/projected/16fbb102-acdd-49fd-a941-9e7ff32b71cc-kube-api-access-27cjl\") pod \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.613208 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-svc\") pod \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.613238 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-config\") pod \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.613331 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-swift-storage-0\") pod \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\" (UID: \"16fbb102-acdd-49fd-a941-9e7ff32b71cc\") " Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.620837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16fbb102-acdd-49fd-a941-9e7ff32b71cc-kube-api-access-27cjl" (OuterVolumeSpecName: "kube-api-access-27cjl") pod "16fbb102-acdd-49fd-a941-9e7ff32b71cc" (UID: "16fbb102-acdd-49fd-a941-9e7ff32b71cc"). InnerVolumeSpecName "kube-api-access-27cjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.637219 4809 scope.go:117] "RemoveContainer" containerID="84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490" Nov 27 17:41:34 crc kubenswrapper[4809]: E1127 17:41:34.638291 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490\": container with ID starting with 84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490 not found: ID does not exist" containerID="84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.638330 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490"} err="failed to get container status \"84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490\": rpc error: code = NotFound desc = could not find container \"84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490\": container with ID starting with 84749f00f14d188418abac968d658acf019b186dd3de6b5195d1df12edb97490 not found: ID does not exist" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.638355 4809 scope.go:117] "RemoveContainer" containerID="d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd" Nov 27 17:41:34 crc kubenswrapper[4809]: E1127 17:41:34.638698 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd\": container with ID starting with d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd not found: ID does not exist" containerID="d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.638816 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd"} err="failed to get container status \"d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd\": rpc error: code = NotFound desc = could not find container \"d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd\": container with ID starting with d197a230e425e8e385c0635bd78dc2715c09633fe93f4b85c6b030ec247b70dd not found: ID does not exist" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.693754 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-config" (OuterVolumeSpecName: "config") pod "16fbb102-acdd-49fd-a941-9e7ff32b71cc" (UID: "16fbb102-acdd-49fd-a941-9e7ff32b71cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.695433 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "16fbb102-acdd-49fd-a941-9e7ff32b71cc" (UID: "16fbb102-acdd-49fd-a941-9e7ff32b71cc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.699302 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "16fbb102-acdd-49fd-a941-9e7ff32b71cc" (UID: "16fbb102-acdd-49fd-a941-9e7ff32b71cc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.702190 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "16fbb102-acdd-49fd-a941-9e7ff32b71cc" (UID: "16fbb102-acdd-49fd-a941-9e7ff32b71cc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.709378 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "16fbb102-acdd-49fd-a941-9e7ff32b71cc" (UID: "16fbb102-acdd-49fd-a941-9e7ff32b71cc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.715966 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.715998 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.716009 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.716021 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.716034 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16fbb102-acdd-49fd-a941-9e7ff32b71cc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.716045 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27cjl\" (UniqueName: \"kubernetes.io/projected/16fbb102-acdd-49fd-a941-9e7ff32b71cc-kube-api-access-27cjl\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.875162 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xtlg4"] Nov 27 17:41:34 crc kubenswrapper[4809]: I1127 17:41:34.888396 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xtlg4"] Nov 27 17:41:35 crc kubenswrapper[4809]: I1127 17:41:35.469300 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16fbb102-acdd-49fd-a941-9e7ff32b71cc" path="/var/lib/kubelet/pods/16fbb102-acdd-49fd-a941-9e7ff32b71cc/volumes" Nov 27 17:41:35 crc kubenswrapper[4809]: I1127 17:41:35.549675 4809 generic.go:334] "Generic (PLEG): container finished" podID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerID="3b10e9f2ef60c91e673f65348b8a867d006b7716d7933446f09153b94268b578" exitCode=0 Nov 27 17:41:35 crc kubenswrapper[4809]: I1127 17:41:35.549707 4809 generic.go:334] "Generic (PLEG): container finished" podID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerID="8b49173719314457e7899b6f9cf63a1bd5e7c69dac6c00349c1d4d7788eb3d44" exitCode=2 Nov 27 17:41:35 crc kubenswrapper[4809]: I1127 17:41:35.549714 4809 generic.go:334] "Generic (PLEG): container finished" podID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerID="73f9da341bfe10da04849f02b3fa14b390b234203c7af032dc113e2521b6cb8a" exitCode=0 Nov 27 17:41:35 crc kubenswrapper[4809]: I1127 17:41:35.549763 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerDied","Data":"3b10e9f2ef60c91e673f65348b8a867d006b7716d7933446f09153b94268b578"} Nov 27 17:41:35 crc kubenswrapper[4809]: I1127 17:41:35.549789 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerDied","Data":"8b49173719314457e7899b6f9cf63a1bd5e7c69dac6c00349c1d4d7788eb3d44"} Nov 27 17:41:35 crc kubenswrapper[4809]: I1127 17:41:35.549800 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerDied","Data":"73f9da341bfe10da04849f02b3fa14b390b234203c7af032dc113e2521b6cb8a"} Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.584368 4809 generic.go:334] "Generic (PLEG): container finished" podID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerID="f941ed7cd0538dfe1787e6d20a8f527b30cc4c43cd05b6efe44e48e81d763d61" exitCode=0 Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.585836 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerDied","Data":"f941ed7cd0538dfe1787e6d20a8f527b30cc4c43cd05b6efe44e48e81d763d61"} Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.830203 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.957010 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-combined-ca-bundle\") pod \"8ae721b0-5274-4f53-a9de-4942464bd7a5\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.957213 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-config-data\") pod \"8ae721b0-5274-4f53-a9de-4942464bd7a5\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.957322 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-sg-core-conf-yaml\") pod \"8ae721b0-5274-4f53-a9de-4942464bd7a5\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.958346 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-run-httpd\") pod \"8ae721b0-5274-4f53-a9de-4942464bd7a5\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.958388 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-log-httpd\") pod \"8ae721b0-5274-4f53-a9de-4942464bd7a5\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.958468 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpp8w\" (UniqueName: \"kubernetes.io/projected/8ae721b0-5274-4f53-a9de-4942464bd7a5-kube-api-access-hpp8w\") pod \"8ae721b0-5274-4f53-a9de-4942464bd7a5\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.958538 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-ceilometer-tls-certs\") pod \"8ae721b0-5274-4f53-a9de-4942464bd7a5\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.958576 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-scripts\") pod \"8ae721b0-5274-4f53-a9de-4942464bd7a5\" (UID: \"8ae721b0-5274-4f53-a9de-4942464bd7a5\") " Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.958633 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8ae721b0-5274-4f53-a9de-4942464bd7a5" (UID: "8ae721b0-5274-4f53-a9de-4942464bd7a5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.958833 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8ae721b0-5274-4f53-a9de-4942464bd7a5" (UID: "8ae721b0-5274-4f53-a9de-4942464bd7a5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.959337 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.959364 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae721b0-5274-4f53-a9de-4942464bd7a5-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.963928 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-scripts" (OuterVolumeSpecName: "scripts") pod "8ae721b0-5274-4f53-a9de-4942464bd7a5" (UID: "8ae721b0-5274-4f53-a9de-4942464bd7a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.972954 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ae721b0-5274-4f53-a9de-4942464bd7a5-kube-api-access-hpp8w" (OuterVolumeSpecName: "kube-api-access-hpp8w") pod "8ae721b0-5274-4f53-a9de-4942464bd7a5" (UID: "8ae721b0-5274-4f53-a9de-4942464bd7a5"). InnerVolumeSpecName "kube-api-access-hpp8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:36 crc kubenswrapper[4809]: I1127 17:41:36.989207 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8ae721b0-5274-4f53-a9de-4942464bd7a5" (UID: "8ae721b0-5274-4f53-a9de-4942464bd7a5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.021662 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8ae721b0-5274-4f53-a9de-4942464bd7a5" (UID: "8ae721b0-5274-4f53-a9de-4942464bd7a5"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.052503 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ae721b0-5274-4f53-a9de-4942464bd7a5" (UID: "8ae721b0-5274-4f53-a9de-4942464bd7a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.061428 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpp8w\" (UniqueName: \"kubernetes.io/projected/8ae721b0-5274-4f53-a9de-4942464bd7a5-kube-api-access-hpp8w\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.061458 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.061466 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.061475 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.061483 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.076383 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-config-data" (OuterVolumeSpecName: "config-data") pod "8ae721b0-5274-4f53-a9de-4942464bd7a5" (UID: "8ae721b0-5274-4f53-a9de-4942464bd7a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.163262 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae721b0-5274-4f53-a9de-4942464bd7a5-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.598297 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae721b0-5274-4f53-a9de-4942464bd7a5","Type":"ContainerDied","Data":"4b7394d563dd5bc7a31ee8cc71ebe0915c02e55bd0da75155c29b8892676881c"} Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.598357 4809 scope.go:117] "RemoveContainer" containerID="3b10e9f2ef60c91e673f65348b8a867d006b7716d7933446f09153b94268b578" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.598476 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.627452 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.629508 4809 scope.go:117] "RemoveContainer" containerID="8b49173719314457e7899b6f9cf63a1bd5e7c69dac6c00349c1d4d7788eb3d44" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.644218 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.654853 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:37 crc kubenswrapper[4809]: E1127 17:41:37.655291 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="ceilometer-notification-agent" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655310 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="ceilometer-notification-agent" Nov 27 17:41:37 crc kubenswrapper[4809]: E1127 17:41:37.655322 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fbb102-acdd-49fd-a941-9e7ff32b71cc" containerName="init" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655329 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fbb102-acdd-49fd-a941-9e7ff32b71cc" containerName="init" Nov 27 17:41:37 crc kubenswrapper[4809]: E1127 17:41:37.655351 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="sg-core" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655357 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="sg-core" Nov 27 17:41:37 crc kubenswrapper[4809]: E1127 17:41:37.655370 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="proxy-httpd" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655375 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="proxy-httpd" Nov 27 17:41:37 crc kubenswrapper[4809]: E1127 17:41:37.655382 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fbb102-acdd-49fd-a941-9e7ff32b71cc" containerName="dnsmasq-dns" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655388 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fbb102-acdd-49fd-a941-9e7ff32b71cc" containerName="dnsmasq-dns" Nov 27 17:41:37 crc kubenswrapper[4809]: E1127 17:41:37.655408 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="ceilometer-central-agent" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655414 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="ceilometer-central-agent" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655592 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="16fbb102-acdd-49fd-a941-9e7ff32b71cc" containerName="dnsmasq-dns" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655603 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="proxy-httpd" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655614 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="ceilometer-notification-agent" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655622 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="sg-core" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.655643 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" containerName="ceilometer-central-agent" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.663134 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.665503 4809 scope.go:117] "RemoveContainer" containerID="73f9da341bfe10da04849f02b3fa14b390b234203c7af032dc113e2521b6cb8a" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.665813 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.666501 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.671451 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.706727 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.707892 4809 scope.go:117] "RemoveContainer" containerID="f941ed7cd0538dfe1787e6d20a8f527b30cc4c43cd05b6efe44e48e81d763d61" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.777884 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffjp9\" (UniqueName: \"kubernetes.io/projected/56782df1-096c-47b7-bebb-b0a0712a46cc-kube-api-access-ffjp9\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.777959 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56782df1-096c-47b7-bebb-b0a0712a46cc-log-httpd\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.777985 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.778030 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-scripts\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.778062 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-config-data\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.778128 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.778248 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56782df1-096c-47b7-bebb-b0a0712a46cc-run-httpd\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.778306 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.879970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56782df1-096c-47b7-bebb-b0a0712a46cc-run-httpd\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.880197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.880381 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffjp9\" (UniqueName: \"kubernetes.io/projected/56782df1-096c-47b7-bebb-b0a0712a46cc-kube-api-access-ffjp9\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.880456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56782df1-096c-47b7-bebb-b0a0712a46cc-log-httpd\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.880525 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.881178 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-scripts\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.880766 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56782df1-096c-47b7-bebb-b0a0712a46cc-run-httpd\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.881265 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-config-data\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.881424 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.881115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56782df1-096c-47b7-bebb-b0a0712a46cc-log-httpd\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.897419 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.897886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-scripts\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.898538 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.899282 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-config-data\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.901804 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffjp9\" (UniqueName: \"kubernetes.io/projected/56782df1-096c-47b7-bebb-b0a0712a46cc-kube-api-access-ffjp9\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.901987 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56782df1-096c-47b7-bebb-b0a0712a46cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56782df1-096c-47b7-bebb-b0a0712a46cc\") " pod="openstack/ceilometer-0" Nov 27 17:41:37 crc kubenswrapper[4809]: I1127 17:41:37.994934 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 17:41:38 crc kubenswrapper[4809]: I1127 17:41:38.457681 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 17:41:38 crc kubenswrapper[4809]: I1127 17:41:38.614797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56782df1-096c-47b7-bebb-b0a0712a46cc","Type":"ContainerStarted","Data":"d7f0177f66331138a42b60f38eef3af6650e84c841087f5f82da205f40b18203"} Nov 27 17:41:39 crc kubenswrapper[4809]: I1127 17:41:39.468406 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ae721b0-5274-4f53-a9de-4942464bd7a5" path="/var/lib/kubelet/pods/8ae721b0-5274-4f53-a9de-4942464bd7a5/volumes" Nov 27 17:41:39 crc kubenswrapper[4809]: I1127 17:41:39.625172 4809 generic.go:334] "Generic (PLEG): container finished" podID="79ca60f5-37e1-4870-8282-7bcb50174a9e" containerID="3df4fcce6640f111dad80a2d5cb0be2f02cd5c4ad1125a7dd6e4e1105e60decc" exitCode=0 Nov 27 17:41:39 crc kubenswrapper[4809]: I1127 17:41:39.625384 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rbmz5" event={"ID":"79ca60f5-37e1-4870-8282-7bcb50174a9e","Type":"ContainerDied","Data":"3df4fcce6640f111dad80a2d5cb0be2f02cd5c4ad1125a7dd6e4e1105e60decc"} Nov 27 17:41:39 crc kubenswrapper[4809]: I1127 17:41:39.628082 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56782df1-096c-47b7-bebb-b0a0712a46cc","Type":"ContainerStarted","Data":"a8cccd16ad529a55d937343f6aeb79a6f315c8a07c897378cdd68f9a4c8b6bf9"} Nov 27 17:41:40 crc kubenswrapper[4809]: I1127 17:41:40.643073 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56782df1-096c-47b7-bebb-b0a0712a46cc","Type":"ContainerStarted","Data":"2ff6481981ce4911043e3e1f3be4cd736b30c90ed7b0c990c135bdd031e5234e"} Nov 27 17:41:40 crc kubenswrapper[4809]: I1127 17:41:40.643421 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56782df1-096c-47b7-bebb-b0a0712a46cc","Type":"ContainerStarted","Data":"73ad1e4e2e99d8989a2df6270603c24cf5ef69c914e60a84a86fef6da7ba5082"} Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.043755 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.148466 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgxv9\" (UniqueName: \"kubernetes.io/projected/79ca60f5-37e1-4870-8282-7bcb50174a9e-kube-api-access-pgxv9\") pod \"79ca60f5-37e1-4870-8282-7bcb50174a9e\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.148619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-combined-ca-bundle\") pod \"79ca60f5-37e1-4870-8282-7bcb50174a9e\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.148665 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-config-data\") pod \"79ca60f5-37e1-4870-8282-7bcb50174a9e\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.148689 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-scripts\") pod \"79ca60f5-37e1-4870-8282-7bcb50174a9e\" (UID: \"79ca60f5-37e1-4870-8282-7bcb50174a9e\") " Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.154638 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79ca60f5-37e1-4870-8282-7bcb50174a9e-kube-api-access-pgxv9" (OuterVolumeSpecName: "kube-api-access-pgxv9") pod "79ca60f5-37e1-4870-8282-7bcb50174a9e" (UID: "79ca60f5-37e1-4870-8282-7bcb50174a9e"). InnerVolumeSpecName "kube-api-access-pgxv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.158937 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-scripts" (OuterVolumeSpecName: "scripts") pod "79ca60f5-37e1-4870-8282-7bcb50174a9e" (UID: "79ca60f5-37e1-4870-8282-7bcb50174a9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.178246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79ca60f5-37e1-4870-8282-7bcb50174a9e" (UID: "79ca60f5-37e1-4870-8282-7bcb50174a9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.179437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-config-data" (OuterVolumeSpecName: "config-data") pod "79ca60f5-37e1-4870-8282-7bcb50174a9e" (UID: "79ca60f5-37e1-4870-8282-7bcb50174a9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.251112 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.251142 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgxv9\" (UniqueName: \"kubernetes.io/projected/79ca60f5-37e1-4870-8282-7bcb50174a9e-kube-api-access-pgxv9\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.251153 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.251162 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ca60f5-37e1-4870-8282-7bcb50174a9e-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.653261 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rbmz5" event={"ID":"79ca60f5-37e1-4870-8282-7bcb50174a9e","Type":"ContainerDied","Data":"2434fe16280fe5dea9b343f3d20025a110924f688af6cfd8357f05b2ba9498f0"} Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.653562 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2434fe16280fe5dea9b343f3d20025a110924f688af6cfd8357f05b2ba9498f0" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.653628 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rbmz5" Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.866896 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.867860 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="887efa58-415d-4bcc-b818-09693edd49f3" containerName="nova-api-log" containerID="cri-o://9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826" gracePeriod=30 Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.867925 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="887efa58-415d-4bcc-b818-09693edd49f3" containerName="nova-api-api" containerID="cri-o://38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13" gracePeriod=30 Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.897271 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.897512 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c6118e4d-990e-42f3-8ab2-584e7b29e3b9" containerName="nova-scheduler-scheduler" containerID="cri-o://875f2953176d32da50aab1ff7340959330701d3059b1571068e337680dc9dad8" gracePeriod=30 Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.917373 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.917672 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-log" containerID="cri-o://3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3" gracePeriod=30 Nov 27 17:41:41 crc kubenswrapper[4809]: I1127 17:41:41.917929 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-metadata" containerID="cri-o://96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107" gracePeriod=30 Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.618150 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.667625 4809 generic.go:334] "Generic (PLEG): container finished" podID="887efa58-415d-4bcc-b818-09693edd49f3" containerID="38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13" exitCode=0 Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.667656 4809 generic.go:334] "Generic (PLEG): container finished" podID="887efa58-415d-4bcc-b818-09693edd49f3" containerID="9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826" exitCode=143 Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.667694 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.667725 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"887efa58-415d-4bcc-b818-09693edd49f3","Type":"ContainerDied","Data":"38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13"} Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.667808 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"887efa58-415d-4bcc-b818-09693edd49f3","Type":"ContainerDied","Data":"9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826"} Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.667825 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"887efa58-415d-4bcc-b818-09693edd49f3","Type":"ContainerDied","Data":"208f30a77604ed58b3a65b5262318cd40db82dae6bec058081f9df939c9f1643"} Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.667844 4809 scope.go:117] "RemoveContainer" containerID="38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.672911 4809 generic.go:334] "Generic (PLEG): container finished" podID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerID="3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3" exitCode=143 Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.672986 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8232a1f9-b58b-4e35-87ec-5d14431c573a","Type":"ContainerDied","Data":"3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3"} Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.695846 4809 scope.go:117] "RemoveContainer" containerID="9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.719663 4809 scope.go:117] "RemoveContainer" containerID="38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13" Nov 27 17:41:42 crc kubenswrapper[4809]: E1127 17:41:42.720142 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13\": container with ID starting with 38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13 not found: ID does not exist" containerID="38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.720200 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13"} err="failed to get container status \"38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13\": rpc error: code = NotFound desc = could not find container \"38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13\": container with ID starting with 38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13 not found: ID does not exist" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.720226 4809 scope.go:117] "RemoveContainer" containerID="9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826" Nov 27 17:41:42 crc kubenswrapper[4809]: E1127 17:41:42.721083 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826\": container with ID starting with 9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826 not found: ID does not exist" containerID="9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.721120 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826"} err="failed to get container status \"9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826\": rpc error: code = NotFound desc = could not find container \"9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826\": container with ID starting with 9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826 not found: ID does not exist" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.721140 4809 scope.go:117] "RemoveContainer" containerID="38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.721397 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13"} err="failed to get container status \"38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13\": rpc error: code = NotFound desc = could not find container \"38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13\": container with ID starting with 38f82c41bd5ece8fbb262c9d2a02ad06ceea2e06e4f4eb2d2baa7e0491614d13 not found: ID does not exist" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.721415 4809 scope.go:117] "RemoveContainer" containerID="9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.721660 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826"} err="failed to get container status \"9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826\": rpc error: code = NotFound desc = could not find container \"9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826\": container with ID starting with 9925e40bf1e5a2d34c8c803ec42f516a37160519f96297e747ebba230f35a826 not found: ID does not exist" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.783726 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr6mc\" (UniqueName: \"kubernetes.io/projected/887efa58-415d-4bcc-b818-09693edd49f3-kube-api-access-qr6mc\") pod \"887efa58-415d-4bcc-b818-09693edd49f3\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.783809 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-config-data\") pod \"887efa58-415d-4bcc-b818-09693edd49f3\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.783904 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-public-tls-certs\") pod \"887efa58-415d-4bcc-b818-09693edd49f3\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.783952 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/887efa58-415d-4bcc-b818-09693edd49f3-logs\") pod \"887efa58-415d-4bcc-b818-09693edd49f3\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.783990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-internal-tls-certs\") pod \"887efa58-415d-4bcc-b818-09693edd49f3\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.784073 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-combined-ca-bundle\") pod \"887efa58-415d-4bcc-b818-09693edd49f3\" (UID: \"887efa58-415d-4bcc-b818-09693edd49f3\") " Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.784471 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/887efa58-415d-4bcc-b818-09693edd49f3-logs" (OuterVolumeSpecName: "logs") pod "887efa58-415d-4bcc-b818-09693edd49f3" (UID: "887efa58-415d-4bcc-b818-09693edd49f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.791688 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/887efa58-415d-4bcc-b818-09693edd49f3-kube-api-access-qr6mc" (OuterVolumeSpecName: "kube-api-access-qr6mc") pod "887efa58-415d-4bcc-b818-09693edd49f3" (UID: "887efa58-415d-4bcc-b818-09693edd49f3"). InnerVolumeSpecName "kube-api-access-qr6mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.816766 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-config-data" (OuterVolumeSpecName: "config-data") pod "887efa58-415d-4bcc-b818-09693edd49f3" (UID: "887efa58-415d-4bcc-b818-09693edd49f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.817634 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "887efa58-415d-4bcc-b818-09693edd49f3" (UID: "887efa58-415d-4bcc-b818-09693edd49f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.839295 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "887efa58-415d-4bcc-b818-09693edd49f3" (UID: "887efa58-415d-4bcc-b818-09693edd49f3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.843328 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "887efa58-415d-4bcc-b818-09693edd49f3" (UID: "887efa58-415d-4bcc-b818-09693edd49f3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.886139 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.886191 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr6mc\" (UniqueName: \"kubernetes.io/projected/887efa58-415d-4bcc-b818-09693edd49f3-kube-api-access-qr6mc\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.886204 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.886215 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.886225 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/887efa58-415d-4bcc-b818-09693edd49f3-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:42 crc kubenswrapper[4809]: I1127 17:41:42.886238 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/887efa58-415d-4bcc-b818-09693edd49f3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.067294 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.081380 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.091852 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:43 crc kubenswrapper[4809]: E1127 17:41:43.092286 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ca60f5-37e1-4870-8282-7bcb50174a9e" containerName="nova-manage" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.092309 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ca60f5-37e1-4870-8282-7bcb50174a9e" containerName="nova-manage" Nov 27 17:41:43 crc kubenswrapper[4809]: E1127 17:41:43.092326 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887efa58-415d-4bcc-b818-09693edd49f3" containerName="nova-api-log" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.092335 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="887efa58-415d-4bcc-b818-09693edd49f3" containerName="nova-api-log" Nov 27 17:41:43 crc kubenswrapper[4809]: E1127 17:41:43.092372 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887efa58-415d-4bcc-b818-09693edd49f3" containerName="nova-api-api" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.092380 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="887efa58-415d-4bcc-b818-09693edd49f3" containerName="nova-api-api" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.092599 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="887efa58-415d-4bcc-b818-09693edd49f3" containerName="nova-api-log" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.092652 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="887efa58-415d-4bcc-b818-09693edd49f3" containerName="nova-api-api" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.092680 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ca60f5-37e1-4870-8282-7bcb50174a9e" containerName="nova-manage" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.093724 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.096679 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.096964 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.097133 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.126630 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.191321 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwcb2\" (UniqueName: \"kubernetes.io/projected/db48eb93-0220-4661-8de6-23aa1e43ca94-kube-api-access-hwcb2\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.191426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-config-data\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.191683 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-internal-tls-certs\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.192005 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-public-tls-certs\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.192171 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.192248 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db48eb93-0220-4661-8de6-23aa1e43ca94-logs\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.293644 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwcb2\" (UniqueName: \"kubernetes.io/projected/db48eb93-0220-4661-8de6-23aa1e43ca94-kube-api-access-hwcb2\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.294062 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-config-data\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.294146 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-internal-tls-certs\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.294191 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-public-tls-certs\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.294232 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.294263 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db48eb93-0220-4661-8de6-23aa1e43ca94-logs\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.294815 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db48eb93-0220-4661-8de6-23aa1e43ca94-logs\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.297859 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-internal-tls-certs\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.299258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-public-tls-certs\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.300673 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.310089 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db48eb93-0220-4661-8de6-23aa1e43ca94-config-data\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.313446 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwcb2\" (UniqueName: \"kubernetes.io/projected/db48eb93-0220-4661-8de6-23aa1e43ca94-kube-api-access-hwcb2\") pod \"nova-api-0\" (UID: \"db48eb93-0220-4661-8de6-23aa1e43ca94\") " pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.410441 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.472922 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="887efa58-415d-4bcc-b818-09693edd49f3" path="/var/lib/kubelet/pods/887efa58-415d-4bcc-b818-09693edd49f3/volumes" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.692106 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56782df1-096c-47b7-bebb-b0a0712a46cc","Type":"ContainerStarted","Data":"0bbc3286a9d174374d4a02c53f52c0ce8584d936b99b8c40f3cfe127b8540c6c"} Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.692425 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.718295 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9506783030000001 podStartE2EDuration="6.718273766s" podCreationTimestamp="2025-11-27 17:41:37 +0000 UTC" firstStartedPulling="2025-11-27 17:41:38.464604728 +0000 UTC m=+1933.737062080" lastFinishedPulling="2025-11-27 17:41:43.232200191 +0000 UTC m=+1938.504657543" observedRunningTime="2025-11-27 17:41:43.711128511 +0000 UTC m=+1938.983585863" watchObservedRunningTime="2025-11-27 17:41:43.718273766 +0000 UTC m=+1938.990731138" Nov 27 17:41:43 crc kubenswrapper[4809]: I1127 17:41:43.853472 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 17:41:43 crc kubenswrapper[4809]: W1127 17:41:43.853792 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb48eb93_0220_4661_8de6_23aa1e43ca94.slice/crio-495c4ff9278777032804beec7d9247fa80ba0ad66f360e600d161bb9bbe1aadd WatchSource:0}: Error finding container 495c4ff9278777032804beec7d9247fa80ba0ad66f360e600d161bb9bbe1aadd: Status 404 returned error can't find the container with id 495c4ff9278777032804beec7d9247fa80ba0ad66f360e600d161bb9bbe1aadd Nov 27 17:41:44 crc kubenswrapper[4809]: I1127 17:41:44.702842 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"db48eb93-0220-4661-8de6-23aa1e43ca94","Type":"ContainerStarted","Data":"0edf3e2a0fb039857ea731ea0a2232f2eb9ecaa485f9230044a40eecbf5e9951"} Nov 27 17:41:44 crc kubenswrapper[4809]: I1127 17:41:44.703186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"db48eb93-0220-4661-8de6-23aa1e43ca94","Type":"ContainerStarted","Data":"7b41279f5c453f0670291422feab6fc4347ce9fd91a04faf220f7563493dae16"} Nov 27 17:41:44 crc kubenswrapper[4809]: I1127 17:41:44.703203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"db48eb93-0220-4661-8de6-23aa1e43ca94","Type":"ContainerStarted","Data":"495c4ff9278777032804beec7d9247fa80ba0ad66f360e600d161bb9bbe1aadd"} Nov 27 17:41:44 crc kubenswrapper[4809]: I1127 17:41:44.722185 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.7221672030000001 podStartE2EDuration="1.722167203s" podCreationTimestamp="2025-11-27 17:41:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:44.719011108 +0000 UTC m=+1939.991468460" watchObservedRunningTime="2025-11-27 17:41:44.722167203 +0000 UTC m=+1939.994624555" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.054592 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": read tcp 10.217.0.2:51258->10.217.0.207:8775: read: connection reset by peer" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.054651 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": read tcp 10.217.0.2:51260->10.217.0.207:8775: read: connection reset by peer" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.643397 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.718224 4809 generic.go:334] "Generic (PLEG): container finished" podID="c6118e4d-990e-42f3-8ab2-584e7b29e3b9" containerID="875f2953176d32da50aab1ff7340959330701d3059b1571068e337680dc9dad8" exitCode=0 Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.718341 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6118e4d-990e-42f3-8ab2-584e7b29e3b9","Type":"ContainerDied","Data":"875f2953176d32da50aab1ff7340959330701d3059b1571068e337680dc9dad8"} Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.721017 4809 generic.go:334] "Generic (PLEG): container finished" podID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerID="96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107" exitCode=0 Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.721854 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8232a1f9-b58b-4e35-87ec-5d14431c573a","Type":"ContainerDied","Data":"96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107"} Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.721896 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8232a1f9-b58b-4e35-87ec-5d14431c573a","Type":"ContainerDied","Data":"873de81636e9b7a52dd578be7e11a04d11fe526947e312fd91083d85848c98d2"} Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.721895 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.722500 4809 scope.go:117] "RemoveContainer" containerID="96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.743992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwvcb\" (UniqueName: \"kubernetes.io/projected/8232a1f9-b58b-4e35-87ec-5d14431c573a-kube-api-access-fwvcb\") pod \"8232a1f9-b58b-4e35-87ec-5d14431c573a\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.744141 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-combined-ca-bundle\") pod \"8232a1f9-b58b-4e35-87ec-5d14431c573a\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.744327 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8232a1f9-b58b-4e35-87ec-5d14431c573a-logs\") pod \"8232a1f9-b58b-4e35-87ec-5d14431c573a\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.744643 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-config-data\") pod \"8232a1f9-b58b-4e35-87ec-5d14431c573a\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.744682 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-nova-metadata-tls-certs\") pod \"8232a1f9-b58b-4e35-87ec-5d14431c573a\" (UID: \"8232a1f9-b58b-4e35-87ec-5d14431c573a\") " Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.746620 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8232a1f9-b58b-4e35-87ec-5d14431c573a-logs" (OuterVolumeSpecName: "logs") pod "8232a1f9-b58b-4e35-87ec-5d14431c573a" (UID: "8232a1f9-b58b-4e35-87ec-5d14431c573a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.753512 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8232a1f9-b58b-4e35-87ec-5d14431c573a-kube-api-access-fwvcb" (OuterVolumeSpecName: "kube-api-access-fwvcb") pod "8232a1f9-b58b-4e35-87ec-5d14431c573a" (UID: "8232a1f9-b58b-4e35-87ec-5d14431c573a"). InnerVolumeSpecName "kube-api-access-fwvcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.755495 4809 scope.go:117] "RemoveContainer" containerID="3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.777665 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-config-data" (OuterVolumeSpecName: "config-data") pod "8232a1f9-b58b-4e35-87ec-5d14431c573a" (UID: "8232a1f9-b58b-4e35-87ec-5d14431c573a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.783974 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8232a1f9-b58b-4e35-87ec-5d14431c573a" (UID: "8232a1f9-b58b-4e35-87ec-5d14431c573a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.810492 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8232a1f9-b58b-4e35-87ec-5d14431c573a" (UID: "8232a1f9-b58b-4e35-87ec-5d14431c573a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.848550 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.848581 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.848595 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwvcb\" (UniqueName: \"kubernetes.io/projected/8232a1f9-b58b-4e35-87ec-5d14431c573a-kube-api-access-fwvcb\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.848607 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8232a1f9-b58b-4e35-87ec-5d14431c573a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.848619 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8232a1f9-b58b-4e35-87ec-5d14431c573a-logs\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.875429 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.882895 4809 scope.go:117] "RemoveContainer" containerID="96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107" Nov 27 17:41:45 crc kubenswrapper[4809]: E1127 17:41:45.884914 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107\": container with ID starting with 96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107 not found: ID does not exist" containerID="96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.884966 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107"} err="failed to get container status \"96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107\": rpc error: code = NotFound desc = could not find container \"96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107\": container with ID starting with 96cd630ed13609159a52ef0d0016b6c0ea210909535960124d204f6c701fc107 not found: ID does not exist" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.885000 4809 scope.go:117] "RemoveContainer" containerID="3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3" Nov 27 17:41:45 crc kubenswrapper[4809]: E1127 17:41:45.885456 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3\": container with ID starting with 3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3 not found: ID does not exist" containerID="3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.885498 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3"} err="failed to get container status \"3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3\": rpc error: code = NotFound desc = could not find container \"3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3\": container with ID starting with 3caed064f295c4d0bb4d5fd921a812446084ff38d2d08f342cfeafa3ae18dbe3 not found: ID does not exist" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.949642 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxn85\" (UniqueName: \"kubernetes.io/projected/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-kube-api-access-dxn85\") pod \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.949841 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-combined-ca-bundle\") pod \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.949951 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-config-data\") pod \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\" (UID: \"c6118e4d-990e-42f3-8ab2-584e7b29e3b9\") " Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.953058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-kube-api-access-dxn85" (OuterVolumeSpecName: "kube-api-access-dxn85") pod "c6118e4d-990e-42f3-8ab2-584e7b29e3b9" (UID: "c6118e4d-990e-42f3-8ab2-584e7b29e3b9"). InnerVolumeSpecName "kube-api-access-dxn85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.978916 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-config-data" (OuterVolumeSpecName: "config-data") pod "c6118e4d-990e-42f3-8ab2-584e7b29e3b9" (UID: "c6118e4d-990e-42f3-8ab2-584e7b29e3b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:45 crc kubenswrapper[4809]: I1127 17:41:45.987352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6118e4d-990e-42f3-8ab2-584e7b29e3b9" (UID: "c6118e4d-990e-42f3-8ab2-584e7b29e3b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.051876 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxn85\" (UniqueName: \"kubernetes.io/projected/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-kube-api-access-dxn85\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.051960 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.051971 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6118e4d-990e-42f3-8ab2-584e7b29e3b9-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.074454 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.088796 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.104925 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:41:46 crc kubenswrapper[4809]: E1127 17:41:46.105453 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-metadata" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.105480 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-metadata" Nov 27 17:41:46 crc kubenswrapper[4809]: E1127 17:41:46.105496 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-log" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.105506 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-log" Nov 27 17:41:46 crc kubenswrapper[4809]: E1127 17:41:46.105549 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6118e4d-990e-42f3-8ab2-584e7b29e3b9" containerName="nova-scheduler-scheduler" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.105558 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6118e4d-990e-42f3-8ab2-584e7b29e3b9" containerName="nova-scheduler-scheduler" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.105814 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6118e4d-990e-42f3-8ab2-584e7b29e3b9" containerName="nova-scheduler-scheduler" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.105846 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-metadata" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.105871 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" containerName="nova-metadata-log" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.107384 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.111828 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.112103 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.132414 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.255211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65be15f8-d10f-4079-b79a-eb92d2cdc432-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.255321 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65be15f8-d10f-4079-b79a-eb92d2cdc432-config-data\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.255345 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65be15f8-d10f-4079-b79a-eb92d2cdc432-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.255547 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqd5h\" (UniqueName: \"kubernetes.io/projected/65be15f8-d10f-4079-b79a-eb92d2cdc432-kube-api-access-tqd5h\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.255692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65be15f8-d10f-4079-b79a-eb92d2cdc432-logs\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.357874 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65be15f8-d10f-4079-b79a-eb92d2cdc432-config-data\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.357933 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65be15f8-d10f-4079-b79a-eb92d2cdc432-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.358005 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqd5h\" (UniqueName: \"kubernetes.io/projected/65be15f8-d10f-4079-b79a-eb92d2cdc432-kube-api-access-tqd5h\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.358071 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65be15f8-d10f-4079-b79a-eb92d2cdc432-logs\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.358169 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65be15f8-d10f-4079-b79a-eb92d2cdc432-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.358502 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65be15f8-d10f-4079-b79a-eb92d2cdc432-logs\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.362949 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65be15f8-d10f-4079-b79a-eb92d2cdc432-config-data\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.363440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65be15f8-d10f-4079-b79a-eb92d2cdc432-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.364131 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65be15f8-d10f-4079-b79a-eb92d2cdc432-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.405732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqd5h\" (UniqueName: \"kubernetes.io/projected/65be15f8-d10f-4079-b79a-eb92d2cdc432-kube-api-access-tqd5h\") pod \"nova-metadata-0\" (UID: \"65be15f8-d10f-4079-b79a-eb92d2cdc432\") " pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.424666 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.737338 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6118e4d-990e-42f3-8ab2-584e7b29e3b9","Type":"ContainerDied","Data":"f8ed8dbe1dcb084bb5b3c9d77d343cee0f27fba6821d7b45672c9f82a93f4e5a"} Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.737879 4809 scope.go:117] "RemoveContainer" containerID="875f2953176d32da50aab1ff7340959330701d3059b1571068e337680dc9dad8" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.737408 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.787723 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.800300 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.819373 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.820794 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.823848 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.830759 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.900354 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 17:41:46 crc kubenswrapper[4809]: W1127 17:41:46.901420 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65be15f8_d10f_4079_b79a_eb92d2cdc432.slice/crio-2ed5599b0d88c91dd6a0881a459d0cc08603ef4f2919a50c03be28e79dab56c8 WatchSource:0}: Error finding container 2ed5599b0d88c91dd6a0881a459d0cc08603ef4f2919a50c03be28e79dab56c8: Status 404 returned error can't find the container with id 2ed5599b0d88c91dd6a0881a459d0cc08603ef4f2919a50c03be28e79dab56c8 Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.976514 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5bbe948-59f7-488a-b62f-ea8ead20541f-config-data\") pod \"nova-scheduler-0\" (UID: \"f5bbe948-59f7-488a-b62f-ea8ead20541f\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.977119 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-987z7\" (UniqueName: \"kubernetes.io/projected/f5bbe948-59f7-488a-b62f-ea8ead20541f-kube-api-access-987z7\") pod \"nova-scheduler-0\" (UID: \"f5bbe948-59f7-488a-b62f-ea8ead20541f\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:46 crc kubenswrapper[4809]: I1127 17:41:46.977167 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5bbe948-59f7-488a-b62f-ea8ead20541f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f5bbe948-59f7-488a-b62f-ea8ead20541f\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.078844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-987z7\" (UniqueName: \"kubernetes.io/projected/f5bbe948-59f7-488a-b62f-ea8ead20541f-kube-api-access-987z7\") pod \"nova-scheduler-0\" (UID: \"f5bbe948-59f7-488a-b62f-ea8ead20541f\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.078898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5bbe948-59f7-488a-b62f-ea8ead20541f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f5bbe948-59f7-488a-b62f-ea8ead20541f\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.078928 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5bbe948-59f7-488a-b62f-ea8ead20541f-config-data\") pod \"nova-scheduler-0\" (UID: \"f5bbe948-59f7-488a-b62f-ea8ead20541f\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.082379 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5bbe948-59f7-488a-b62f-ea8ead20541f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f5bbe948-59f7-488a-b62f-ea8ead20541f\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.083674 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5bbe948-59f7-488a-b62f-ea8ead20541f-config-data\") pod \"nova-scheduler-0\" (UID: \"f5bbe948-59f7-488a-b62f-ea8ead20541f\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.098293 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-987z7\" (UniqueName: \"kubernetes.io/projected/f5bbe948-59f7-488a-b62f-ea8ead20541f-kube-api-access-987z7\") pod \"nova-scheduler-0\" (UID: \"f5bbe948-59f7-488a-b62f-ea8ead20541f\") " pod="openstack/nova-scheduler-0" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.145786 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.471840 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8232a1f9-b58b-4e35-87ec-5d14431c573a" path="/var/lib/kubelet/pods/8232a1f9-b58b-4e35-87ec-5d14431c573a/volumes" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.472809 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6118e4d-990e-42f3-8ab2-584e7b29e3b9" path="/var/lib/kubelet/pods/c6118e4d-990e-42f3-8ab2-584e7b29e3b9/volumes" Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.597397 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 17:41:47 crc kubenswrapper[4809]: W1127 17:41:47.598829 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5bbe948_59f7_488a_b62f_ea8ead20541f.slice/crio-847ce0323e380ccb0912bfbb5df318017acd17bdd6ba87ae0bbca4c8c520d004 WatchSource:0}: Error finding container 847ce0323e380ccb0912bfbb5df318017acd17bdd6ba87ae0bbca4c8c520d004: Status 404 returned error can't find the container with id 847ce0323e380ccb0912bfbb5df318017acd17bdd6ba87ae0bbca4c8c520d004 Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.749452 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f5bbe948-59f7-488a-b62f-ea8ead20541f","Type":"ContainerStarted","Data":"847ce0323e380ccb0912bfbb5df318017acd17bdd6ba87ae0bbca4c8c520d004"} Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.752869 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65be15f8-d10f-4079-b79a-eb92d2cdc432","Type":"ContainerStarted","Data":"4c434f6ce84fbe7caf1199d8f7fbe7c73d215fc8fd4759208e00e120546632ca"} Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.752912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65be15f8-d10f-4079-b79a-eb92d2cdc432","Type":"ContainerStarted","Data":"8fc01c26a2341d88ecad2fb33f51126f439d13e1ad32b27ac33c63cc6e16d506"} Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.752926 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65be15f8-d10f-4079-b79a-eb92d2cdc432","Type":"ContainerStarted","Data":"2ed5599b0d88c91dd6a0881a459d0cc08603ef4f2919a50c03be28e79dab56c8"} Nov 27 17:41:47 crc kubenswrapper[4809]: I1127 17:41:47.777044 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.777024748 podStartE2EDuration="1.777024748s" podCreationTimestamp="2025-11-27 17:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:47.775331512 +0000 UTC m=+1943.047788884" watchObservedRunningTime="2025-11-27 17:41:47.777024748 +0000 UTC m=+1943.049482120" Nov 27 17:41:48 crc kubenswrapper[4809]: I1127 17:41:48.764927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f5bbe948-59f7-488a-b62f-ea8ead20541f","Type":"ContainerStarted","Data":"e105ef014c3e44212e16a1c40a19025abc8485b88dfc246be66c377a8996d522"} Nov 27 17:41:48 crc kubenswrapper[4809]: I1127 17:41:48.787521 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7874967550000003 podStartE2EDuration="2.787496755s" podCreationTimestamp="2025-11-27 17:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:41:48.778313855 +0000 UTC m=+1944.050771207" watchObservedRunningTime="2025-11-27 17:41:48.787496755 +0000 UTC m=+1944.059954127" Nov 27 17:41:49 crc kubenswrapper[4809]: I1127 17:41:49.458921 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:41:49 crc kubenswrapper[4809]: E1127 17:41:49.459128 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:41:51 crc kubenswrapper[4809]: I1127 17:41:51.425699 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 17:41:51 crc kubenswrapper[4809]: I1127 17:41:51.426173 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 17:41:52 crc kubenswrapper[4809]: I1127 17:41:52.146907 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 27 17:41:53 crc kubenswrapper[4809]: I1127 17:41:53.411819 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 17:41:53 crc kubenswrapper[4809]: I1127 17:41:53.411869 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 17:41:54 crc kubenswrapper[4809]: I1127 17:41:54.428905 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="db48eb93-0220-4661-8de6-23aa1e43ca94" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.218:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 17:41:54 crc kubenswrapper[4809]: I1127 17:41:54.428913 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="db48eb93-0220-4661-8de6-23aa1e43ca94" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.218:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 17:41:56 crc kubenswrapper[4809]: I1127 17:41:56.426108 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 17:41:56 crc kubenswrapper[4809]: I1127 17:41:56.426888 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 17:41:57 crc kubenswrapper[4809]: I1127 17:41:57.147125 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 27 17:41:57 crc kubenswrapper[4809]: I1127 17:41:57.174873 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 27 17:41:57 crc kubenswrapper[4809]: I1127 17:41:57.441925 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="65be15f8-d10f-4079-b79a-eb92d2cdc432" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 17:41:57 crc kubenswrapper[4809]: I1127 17:41:57.441923 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="65be15f8-d10f-4079-b79a-eb92d2cdc432" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 17:41:57 crc kubenswrapper[4809]: I1127 17:41:57.875314 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 27 17:42:01 crc kubenswrapper[4809]: I1127 17:42:01.501519 4809 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod1895833b-2f96-438e-8ab2-85b1d9752b87"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod1895833b-2f96-438e-8ab2-85b1d9752b87] : Timed out while waiting for systemd to remove kubepods-besteffort-pod1895833b_2f96_438e_8ab2_85b1d9752b87.slice" Nov 27 17:42:02 crc kubenswrapper[4809]: I1127 17:42:02.458686 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:42:02 crc kubenswrapper[4809]: E1127 17:42:02.459141 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:42:03 crc kubenswrapper[4809]: I1127 17:42:03.418833 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 17:42:03 crc kubenswrapper[4809]: I1127 17:42:03.419574 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 17:42:03 crc kubenswrapper[4809]: I1127 17:42:03.420566 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 17:42:03 crc kubenswrapper[4809]: I1127 17:42:03.425316 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 17:42:03 crc kubenswrapper[4809]: I1127 17:42:03.918208 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 17:42:03 crc kubenswrapper[4809]: I1127 17:42:03.930789 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 17:42:06 crc kubenswrapper[4809]: I1127 17:42:06.433177 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 17:42:06 crc kubenswrapper[4809]: I1127 17:42:06.434306 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 17:42:06 crc kubenswrapper[4809]: I1127 17:42:06.439357 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 17:42:06 crc kubenswrapper[4809]: I1127 17:42:06.952021 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 17:42:08 crc kubenswrapper[4809]: I1127 17:42:08.015789 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 27 17:42:15 crc kubenswrapper[4809]: I1127 17:42:15.464540 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:42:15 crc kubenswrapper[4809]: E1127 17:42:15.465320 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:42:17 crc kubenswrapper[4809]: I1127 17:42:17.512085 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 17:42:18 crc kubenswrapper[4809]: I1127 17:42:18.899671 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 17:42:21 crc kubenswrapper[4809]: I1127 17:42:21.308756 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="01288244-f670-47a4-ac43-180191d7f331" containerName="rabbitmq" containerID="cri-o://5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439" gracePeriod=604797 Nov 27 17:42:23 crc kubenswrapper[4809]: I1127 17:42:23.009106 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="f8145e23-a84e-405f-beb6-e27bb2aa1c59" containerName="rabbitmq" containerID="cri-o://9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d" gracePeriod=604796 Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.933208 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986307 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986389 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-plugins-conf\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986440 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-plugins\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986485 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-confd\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986513 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01288244-f670-47a4-ac43-180191d7f331-pod-info\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986578 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-tls\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986595 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01288244-f670-47a4-ac43-180191d7f331-erlang-cookie-secret\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986624 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-config-data\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986641 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-server-conf\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986699 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-erlang-cookie\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.986728 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz6n6\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-kube-api-access-nz6n6\") pod \"01288244-f670-47a4-ac43-180191d7f331\" (UID: \"01288244-f670-47a4-ac43-180191d7f331\") " Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.987328 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.988112 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.991596 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.995902 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01288244-f670-47a4-ac43-180191d7f331-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.995899 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.996119 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-kube-api-access-nz6n6" (OuterVolumeSpecName: "kube-api-access-nz6n6") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "kube-api-access-nz6n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:42:27 crc kubenswrapper[4809]: I1127 17:42:27.997502 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/01288244-f670-47a4-ac43-180191d7f331-pod-info" (OuterVolumeSpecName: "pod-info") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.003903 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.046552 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-config-data" (OuterVolumeSpecName: "config-data") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.073550 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-server-conf" (OuterVolumeSpecName: "server-conf") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089223 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089263 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089273 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089281 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/01288244-f670-47a4-ac43-180191d7f331-pod-info\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089290 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089298 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/01288244-f670-47a4-ac43-180191d7f331-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089307 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089315 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/01288244-f670-47a4-ac43-180191d7f331-server-conf\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089323 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.089331 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz6n6\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-kube-api-access-nz6n6\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.111646 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.140622 4809 generic.go:334] "Generic (PLEG): container finished" podID="01288244-f670-47a4-ac43-180191d7f331" containerID="5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439" exitCode=0 Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.140661 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01288244-f670-47a4-ac43-180191d7f331","Type":"ContainerDied","Data":"5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439"} Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.140684 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"01288244-f670-47a4-ac43-180191d7f331","Type":"ContainerDied","Data":"8f46c1e32fcaae1fb5a8e6dc996aefc3138b25bd3a8696657ba6983c6225d736"} Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.140702 4809 scope.go:117] "RemoveContainer" containerID="5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.140839 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.163958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "01288244-f670-47a4-ac43-180191d7f331" (UID: "01288244-f670-47a4-ac43-180191d7f331"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.166762 4809 scope.go:117] "RemoveContainer" containerID="dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.190981 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.191201 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/01288244-f670-47a4-ac43-180191d7f331-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.193589 4809 scope.go:117] "RemoveContainer" containerID="5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439" Nov 27 17:42:28 crc kubenswrapper[4809]: E1127 17:42:28.194054 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439\": container with ID starting with 5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439 not found: ID does not exist" containerID="5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.194148 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439"} err="failed to get container status \"5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439\": rpc error: code = NotFound desc = could not find container \"5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439\": container with ID starting with 5e0affa1eeee9ae21ba1534992aab92359a2274e8adcc34530dfb2841fc52439 not found: ID does not exist" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.194235 4809 scope.go:117] "RemoveContainer" containerID="dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b" Nov 27 17:42:28 crc kubenswrapper[4809]: E1127 17:42:28.194554 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b\": container with ID starting with dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b not found: ID does not exist" containerID="dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.194598 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b"} err="failed to get container status \"dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b\": rpc error: code = NotFound desc = could not find container \"dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b\": container with ID starting with dcd12dbe45b99b1d3fa3933a2417de03925e0a327fafbe06f9573517b1f3358b not found: ID does not exist" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.459132 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.474461 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.484179 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.512360 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 17:42:28 crc kubenswrapper[4809]: E1127 17:42:28.515501 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01288244-f670-47a4-ac43-180191d7f331" containerName="setup-container" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.515527 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="01288244-f670-47a4-ac43-180191d7f331" containerName="setup-container" Nov 27 17:42:28 crc kubenswrapper[4809]: E1127 17:42:28.515566 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01288244-f670-47a4-ac43-180191d7f331" containerName="rabbitmq" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.515575 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="01288244-f670-47a4-ac43-180191d7f331" containerName="rabbitmq" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.519790 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="01288244-f670-47a4-ac43-180191d7f331" containerName="rabbitmq" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.527596 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.529486 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.541356 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5mnsp" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.541771 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.543379 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.544661 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.545092 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.545371 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.561523 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.597853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62edd287-129b-48b0-af61-d21bd27590d7-config-data\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.597893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/62edd287-129b-48b0-af61-d21bd27590d7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.597918 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/62edd287-129b-48b0-af61-d21bd27590d7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.597935 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.597962 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.597993 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/62edd287-129b-48b0-af61-d21bd27590d7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.598196 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.598302 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b257g\" (UniqueName: \"kubernetes.io/projected/62edd287-129b-48b0-af61-d21bd27590d7-kube-api-access-b257g\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.598348 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/62edd287-129b-48b0-af61-d21bd27590d7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.598463 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.598543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.699508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.699863 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.699899 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62edd287-129b-48b0-af61-d21bd27590d7-config-data\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.699918 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/62edd287-129b-48b0-af61-d21bd27590d7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.699940 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/62edd287-129b-48b0-af61-d21bd27590d7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.699956 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.699990 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.700019 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/62edd287-129b-48b0-af61-d21bd27590d7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.700048 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.700070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b257g\" (UniqueName: \"kubernetes.io/projected/62edd287-129b-48b0-af61-d21bd27590d7-kube-api-access-b257g\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.700087 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/62edd287-129b-48b0-af61-d21bd27590d7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.700477 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.700708 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.700820 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.700850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62edd287-129b-48b0-af61-d21bd27590d7-config-data\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.701175 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/62edd287-129b-48b0-af61-d21bd27590d7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.701527 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/62edd287-129b-48b0-af61-d21bd27590d7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.703642 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/62edd287-129b-48b0-af61-d21bd27590d7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.704239 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.705857 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/62edd287-129b-48b0-af61-d21bd27590d7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.706289 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/62edd287-129b-48b0-af61-d21bd27590d7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.720684 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b257g\" (UniqueName: \"kubernetes.io/projected/62edd287-129b-48b0-af61-d21bd27590d7-kube-api-access-b257g\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.735615 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"62edd287-129b-48b0-af61-d21bd27590d7\") " pod="openstack/rabbitmq-server-0" Nov 27 17:42:28 crc kubenswrapper[4809]: I1127 17:42:28.917450 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 17:42:29 crc kubenswrapper[4809]: I1127 17:42:29.159729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"9a6ef73344f0f4b1da4284fd6ac88d1404b7c5e1410dce4a9042ec288f5d1365"} Nov 27 17:42:29 crc kubenswrapper[4809]: I1127 17:42:29.388967 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 17:42:29 crc kubenswrapper[4809]: W1127 17:42:29.389444 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62edd287_129b_48b0_af61_d21bd27590d7.slice/crio-94b8a01b9fb7da6788fd9dfcdb0e1f1d95aac51543bc1b52f9b093e8741c1eda WatchSource:0}: Error finding container 94b8a01b9fb7da6788fd9dfcdb0e1f1d95aac51543bc1b52f9b093e8741c1eda: Status 404 returned error can't find the container with id 94b8a01b9fb7da6788fd9dfcdb0e1f1d95aac51543bc1b52f9b093e8741c1eda Nov 27 17:42:29 crc kubenswrapper[4809]: I1127 17:42:29.468635 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01288244-f670-47a4-ac43-180191d7f331" path="/var/lib/kubelet/pods/01288244-f670-47a4-ac43-180191d7f331/volumes" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.037472 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133026 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdlbl\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-kube-api-access-sdlbl\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133069 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-config-data\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133115 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133160 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-confd\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133234 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8145e23-a84e-405f-beb6-e27bb2aa1c59-erlang-cookie-secret\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133260 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8145e23-a84e-405f-beb6-e27bb2aa1c59-pod-info\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-plugins\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133389 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-plugins-conf\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133430 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-server-conf\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133461 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-tls\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.133484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-erlang-cookie\") pod \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\" (UID: \"f8145e23-a84e-405f-beb6-e27bb2aa1c59\") " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.135170 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.141017 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.143299 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.148850 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-kube-api-access-sdlbl" (OuterVolumeSpecName: "kube-api-access-sdlbl") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "kube-api-access-sdlbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.151075 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.151470 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.152388 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f8145e23-a84e-405f-beb6-e27bb2aa1c59-pod-info" (OuterVolumeSpecName: "pod-info") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.153015 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.154013 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8145e23-a84e-405f-beb6-e27bb2aa1c59-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.174321 4809 generic.go:334] "Generic (PLEG): container finished" podID="f8145e23-a84e-405f-beb6-e27bb2aa1c59" containerID="9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d" exitCode=0 Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.174904 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.174976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8145e23-a84e-405f-beb6-e27bb2aa1c59","Type":"ContainerDied","Data":"9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d"} Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.175060 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f8145e23-a84e-405f-beb6-e27bb2aa1c59","Type":"ContainerDied","Data":"7b776110c98052a76f6a2f9d39014d8cb0a8abc8d6c4405df4980798e25e8dd7"} Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.175083 4809 scope.go:117] "RemoveContainer" containerID="9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.178418 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62edd287-129b-48b0-af61-d21bd27590d7","Type":"ContainerStarted","Data":"94b8a01b9fb7da6788fd9dfcdb0e1f1d95aac51543bc1b52f9b093e8741c1eda"} Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.183941 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-config-data" (OuterVolumeSpecName: "config-data") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.202647 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-server-conf" (OuterVolumeSpecName: "server-conf") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.243017 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.243376 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-server-conf\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.243389 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.243397 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.243407 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdlbl\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-kube-api-access-sdlbl\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.243416 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8145e23-a84e-405f-beb6-e27bb2aa1c59-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.243437 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.243447 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8145e23-a84e-405f-beb6-e27bb2aa1c59-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.243455 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8145e23-a84e-405f-beb6-e27bb2aa1c59-pod-info\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.333457 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.345313 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.688364 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f8145e23-a84e-405f-beb6-e27bb2aa1c59" (UID: "f8145e23-a84e-405f-beb6-e27bb2aa1c59"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.746624 4809 scope.go:117] "RemoveContainer" containerID="e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.752253 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8145e23-a84e-405f-beb6-e27bb2aa1c59-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.784553 4809 scope.go:117] "RemoveContainer" containerID="9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d" Nov 27 17:42:30 crc kubenswrapper[4809]: E1127 17:42:30.785281 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d\": container with ID starting with 9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d not found: ID does not exist" containerID="9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.785320 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d"} err="failed to get container status \"9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d\": rpc error: code = NotFound desc = could not find container \"9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d\": container with ID starting with 9fdba3a63eb880fdaa734a13f6251c67e03b9661aea5ea9c4b3f35ab8320dd6d not found: ID does not exist" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.785349 4809 scope.go:117] "RemoveContainer" containerID="e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5" Nov 27 17:42:30 crc kubenswrapper[4809]: E1127 17:42:30.785684 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5\": container with ID starting with e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5 not found: ID does not exist" containerID="e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.785712 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5"} err="failed to get container status \"e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5\": rpc error: code = NotFound desc = could not find container \"e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5\": container with ID starting with e682373daeabdfb0723cec0dbfde436963291b1bbd157b62d234ce7aa77feec5 not found: ID does not exist" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.827374 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.840545 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.853279 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 17:42:30 crc kubenswrapper[4809]: E1127 17:42:30.853832 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8145e23-a84e-405f-beb6-e27bb2aa1c59" containerName="setup-container" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.853855 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8145e23-a84e-405f-beb6-e27bb2aa1c59" containerName="setup-container" Nov 27 17:42:30 crc kubenswrapper[4809]: E1127 17:42:30.853888 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8145e23-a84e-405f-beb6-e27bb2aa1c59" containerName="rabbitmq" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.853896 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8145e23-a84e-405f-beb6-e27bb2aa1c59" containerName="rabbitmq" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.854149 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8145e23-a84e-405f-beb6-e27bb2aa1c59" containerName="rabbitmq" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.855489 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.858421 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.858564 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.858588 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6rbpw" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.858840 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.858958 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.859047 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.859308 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.865726 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956345 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3f0cad7c-85ca-4269-be47-555613716799-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956395 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f0cad7c-85ca-4269-be47-555613716799-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956432 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3f0cad7c-85ca-4269-be47-555613716799-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956561 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4hlc\" (UniqueName: \"kubernetes.io/projected/3f0cad7c-85ca-4269-be47-555613716799-kube-api-access-f4hlc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956590 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956611 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956644 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956667 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956687 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3f0cad7c-85ca-4269-be47-555613716799-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:30 crc kubenswrapper[4809]: I1127 17:42:30.956841 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3f0cad7c-85ca-4269-be47-555613716799-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058543 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3f0cad7c-85ca-4269-be47-555613716799-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058644 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3f0cad7c-85ca-4269-be47-555613716799-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f0cad7c-85ca-4269-be47-555613716799-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058721 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058783 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3f0cad7c-85ca-4269-be47-555613716799-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4hlc\" (UniqueName: \"kubernetes.io/projected/3f0cad7c-85ca-4269-be47-555613716799-kube-api-access-f4hlc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058878 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058894 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058944 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.058964 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3f0cad7c-85ca-4269-be47-555613716799-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.059766 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.059828 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f0cad7c-85ca-4269-be47-555613716799-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.059891 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3f0cad7c-85ca-4269-be47-555613716799-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.060044 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.060075 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.060436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3f0cad7c-85ca-4269-be47-555613716799-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.064380 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.064555 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3f0cad7c-85ca-4269-be47-555613716799-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.065213 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3f0cad7c-85ca-4269-be47-555613716799-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.066309 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3f0cad7c-85ca-4269-be47-555613716799-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.076205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4hlc\" (UniqueName: \"kubernetes.io/projected/3f0cad7c-85ca-4269-be47-555613716799-kube-api-access-f4hlc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.094283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3f0cad7c-85ca-4269-be47-555613716799\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.181509 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.188886 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62edd287-129b-48b0-af61-d21bd27590d7","Type":"ContainerStarted","Data":"0428b26a5044722c21b49bf8b09b9c999ee5edac77f0e5e39491eeea99adee29"} Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.470784 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8145e23-a84e-405f-beb6-e27bb2aa1c59" path="/var/lib/kubelet/pods/f8145e23-a84e-405f-beb6-e27bb2aa1c59/volumes" Nov 27 17:42:31 crc kubenswrapper[4809]: I1127 17:42:31.627540 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.201297 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3f0cad7c-85ca-4269-be47-555613716799","Type":"ContainerStarted","Data":"ed0ee033aaf2d48e557c2eb0a6d25accb1d0f87308d3dd74c6c9d4c62f638c0c"} Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.442258 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-wwxf2"] Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.446675 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.452512 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.459952 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-wwxf2"] Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.590829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.590908 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-svc\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.590933 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.590979 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.591175 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-config\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.591363 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4bqv\" (UniqueName: \"kubernetes.io/projected/999823be-7414-421c-8ec1-dfeb6981546a-kube-api-access-x4bqv\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.591472 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.692828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-config\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.692916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4bqv\" (UniqueName: \"kubernetes.io/projected/999823be-7414-421c-8ec1-dfeb6981546a-kube-api-access-x4bqv\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.692976 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.693035 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.693086 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-svc\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.693111 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.693176 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.693943 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-config\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.694034 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.694131 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.694149 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-svc\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.694194 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.694506 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:32 crc kubenswrapper[4809]: I1127 17:42:32.909118 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4bqv\" (UniqueName: \"kubernetes.io/projected/999823be-7414-421c-8ec1-dfeb6981546a-kube-api-access-x4bqv\") pod \"dnsmasq-dns-67b789f86c-wwxf2\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:33 crc kubenswrapper[4809]: I1127 17:42:33.066669 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:33 crc kubenswrapper[4809]: I1127 17:42:33.212868 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3f0cad7c-85ca-4269-be47-555613716799","Type":"ContainerStarted","Data":"adba087f0b808b4f6fab80a9be3fa2d5a09a50e4f0b0614c8d64ab9a4fcac910"} Nov 27 17:42:33 crc kubenswrapper[4809]: I1127 17:42:33.518083 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-wwxf2"] Nov 27 17:42:33 crc kubenswrapper[4809]: W1127 17:42:33.519227 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod999823be_7414_421c_8ec1_dfeb6981546a.slice/crio-376f07d6830140f0e839ee02d2f16ba3e792d81267bb9f40ab8670eca6cb0f22 WatchSource:0}: Error finding container 376f07d6830140f0e839ee02d2f16ba3e792d81267bb9f40ab8670eca6cb0f22: Status 404 returned error can't find the container with id 376f07d6830140f0e839ee02d2f16ba3e792d81267bb9f40ab8670eca6cb0f22 Nov 27 17:42:34 crc kubenswrapper[4809]: I1127 17:42:34.222912 4809 generic.go:334] "Generic (PLEG): container finished" podID="999823be-7414-421c-8ec1-dfeb6981546a" containerID="973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac" exitCode=0 Nov 27 17:42:34 crc kubenswrapper[4809]: I1127 17:42:34.223012 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" event={"ID":"999823be-7414-421c-8ec1-dfeb6981546a","Type":"ContainerDied","Data":"973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac"} Nov 27 17:42:34 crc kubenswrapper[4809]: I1127 17:42:34.223381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" event={"ID":"999823be-7414-421c-8ec1-dfeb6981546a","Type":"ContainerStarted","Data":"376f07d6830140f0e839ee02d2f16ba3e792d81267bb9f40ab8670eca6cb0f22"} Nov 27 17:42:35 crc kubenswrapper[4809]: I1127 17:42:35.234247 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" event={"ID":"999823be-7414-421c-8ec1-dfeb6981546a","Type":"ContainerStarted","Data":"89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a"} Nov 27 17:42:35 crc kubenswrapper[4809]: I1127 17:42:35.234555 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:35 crc kubenswrapper[4809]: I1127 17:42:35.261650 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" podStartSLOduration=3.261628311 podStartE2EDuration="3.261628311s" podCreationTimestamp="2025-11-27 17:42:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:42:35.254084717 +0000 UTC m=+1990.526542069" watchObservedRunningTime="2025-11-27 17:42:35.261628311 +0000 UTC m=+1990.534085663" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.069052 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.146416 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-tcmnm"] Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.146636 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" podUID="14a8fdea-bcf5-49a2-8b88-13875ac05c71" containerName="dnsmasq-dns" containerID="cri-o://487c143437edb8bde10d7fd825cbecf6c0f79773064bb64afa371599825ed8ac" gracePeriod=10 Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.289381 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-zgq88"] Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.291328 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.301070 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-zgq88"] Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.314630 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.314931 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlxcd\" (UniqueName: \"kubernetes.io/projected/0025bfa8-25eb-43cc-825e-c4e299f28d47-kube-api-access-wlxcd\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.314984 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-config\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.315094 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.315303 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.315413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.315441 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.329368 4809 generic.go:334] "Generic (PLEG): container finished" podID="14a8fdea-bcf5-49a2-8b88-13875ac05c71" containerID="487c143437edb8bde10d7fd825cbecf6c0f79773064bb64afa371599825ed8ac" exitCode=0 Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.329435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" event={"ID":"14a8fdea-bcf5-49a2-8b88-13875ac05c71","Type":"ContainerDied","Data":"487c143437edb8bde10d7fd825cbecf6c0f79773064bb64afa371599825ed8ac"} Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.416091 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.416339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.416442 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.416550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.416655 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlxcd\" (UniqueName: \"kubernetes.io/projected/0025bfa8-25eb-43cc-825e-c4e299f28d47-kube-api-access-wlxcd\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.416825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-config\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.416920 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.417048 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.417642 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.417909 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.418075 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-config\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.418075 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.419011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0025bfa8-25eb-43cc-825e-c4e299f28d47-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.465244 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlxcd\" (UniqueName: \"kubernetes.io/projected/0025bfa8-25eb-43cc-825e-c4e299f28d47-kube-api-access-wlxcd\") pod \"dnsmasq-dns-cb6ffcf87-zgq88\" (UID: \"0025bfa8-25eb-43cc-825e-c4e299f28d47\") " pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.640449 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.779526 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.832788 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-swift-storage-0\") pod \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.832867 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-nb\") pod \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.832924 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-sb\") pod \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.832966 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-svc\") pod \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.833026 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-config\") pod \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.833055 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgpl6\" (UniqueName: \"kubernetes.io/projected/14a8fdea-bcf5-49a2-8b88-13875ac05c71-kube-api-access-sgpl6\") pod \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\" (UID: \"14a8fdea-bcf5-49a2-8b88-13875ac05c71\") " Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.852518 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a8fdea-bcf5-49a2-8b88-13875ac05c71-kube-api-access-sgpl6" (OuterVolumeSpecName: "kube-api-access-sgpl6") pod "14a8fdea-bcf5-49a2-8b88-13875ac05c71" (UID: "14a8fdea-bcf5-49a2-8b88-13875ac05c71"). InnerVolumeSpecName "kube-api-access-sgpl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.897942 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-config" (OuterVolumeSpecName: "config") pod "14a8fdea-bcf5-49a2-8b88-13875ac05c71" (UID: "14a8fdea-bcf5-49a2-8b88-13875ac05c71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.898824 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "14a8fdea-bcf5-49a2-8b88-13875ac05c71" (UID: "14a8fdea-bcf5-49a2-8b88-13875ac05c71"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.899396 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14a8fdea-bcf5-49a2-8b88-13875ac05c71" (UID: "14a8fdea-bcf5-49a2-8b88-13875ac05c71"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.908897 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "14a8fdea-bcf5-49a2-8b88-13875ac05c71" (UID: "14a8fdea-bcf5-49a2-8b88-13875ac05c71"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.912116 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "14a8fdea-bcf5-49a2-8b88-13875ac05c71" (UID: "14a8fdea-bcf5-49a2-8b88-13875ac05c71"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.935972 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.936007 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.936019 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.936033 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgpl6\" (UniqueName: \"kubernetes.io/projected/14a8fdea-bcf5-49a2-8b88-13875ac05c71-kube-api-access-sgpl6\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.936049 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:43 crc kubenswrapper[4809]: I1127 17:42:43.936061 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a8fdea-bcf5-49a2-8b88-13875ac05c71-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:44 crc kubenswrapper[4809]: W1127 17:42:44.117444 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0025bfa8_25eb_43cc_825e_c4e299f28d47.slice/crio-ab48ace6575b026d54fa6a8518cc475a5da37151284eed442580e8fb8ecd365f WatchSource:0}: Error finding container ab48ace6575b026d54fa6a8518cc475a5da37151284eed442580e8fb8ecd365f: Status 404 returned error can't find the container with id ab48ace6575b026d54fa6a8518cc475a5da37151284eed442580e8fb8ecd365f Nov 27 17:42:44 crc kubenswrapper[4809]: I1127 17:42:44.118037 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-zgq88"] Nov 27 17:42:44 crc kubenswrapper[4809]: I1127 17:42:44.344427 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" event={"ID":"14a8fdea-bcf5-49a2-8b88-13875ac05c71","Type":"ContainerDied","Data":"953256730a2a0506a772c41863723363afc6f72239d4ce5f07bf7b4c2ab45f50"} Nov 27 17:42:44 crc kubenswrapper[4809]: I1127 17:42:44.344825 4809 scope.go:117] "RemoveContainer" containerID="487c143437edb8bde10d7fd825cbecf6c0f79773064bb64afa371599825ed8ac" Nov 27 17:42:44 crc kubenswrapper[4809]: I1127 17:42:44.344996 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-tcmnm" Nov 27 17:42:44 crc kubenswrapper[4809]: I1127 17:42:44.349332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" event={"ID":"0025bfa8-25eb-43cc-825e-c4e299f28d47","Type":"ContainerStarted","Data":"3c37b69a7240fa76d8897fa2428ca284bf4f53dff92ceaaad827b4cee3f3756c"} Nov 27 17:42:44 crc kubenswrapper[4809]: I1127 17:42:44.349375 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" event={"ID":"0025bfa8-25eb-43cc-825e-c4e299f28d47","Type":"ContainerStarted","Data":"ab48ace6575b026d54fa6a8518cc475a5da37151284eed442580e8fb8ecd365f"} Nov 27 17:42:44 crc kubenswrapper[4809]: I1127 17:42:44.503505 4809 scope.go:117] "RemoveContainer" containerID="d534bd6bf4808e319567a79fb1432f5d93ad40467433da850231a465b186b97a" Nov 27 17:42:44 crc kubenswrapper[4809]: I1127 17:42:44.540119 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-tcmnm"] Nov 27 17:42:44 crc kubenswrapper[4809]: I1127 17:42:44.551075 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-tcmnm"] Nov 27 17:42:45 crc kubenswrapper[4809]: I1127 17:42:45.363330 4809 generic.go:334] "Generic (PLEG): container finished" podID="0025bfa8-25eb-43cc-825e-c4e299f28d47" containerID="3c37b69a7240fa76d8897fa2428ca284bf4f53dff92ceaaad827b4cee3f3756c" exitCode=0 Nov 27 17:42:45 crc kubenswrapper[4809]: I1127 17:42:45.363423 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" event={"ID":"0025bfa8-25eb-43cc-825e-c4e299f28d47","Type":"ContainerDied","Data":"3c37b69a7240fa76d8897fa2428ca284bf4f53dff92ceaaad827b4cee3f3756c"} Nov 27 17:42:45 crc kubenswrapper[4809]: I1127 17:42:45.363864 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:45 crc kubenswrapper[4809]: I1127 17:42:45.363883 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" event={"ID":"0025bfa8-25eb-43cc-825e-c4e299f28d47","Type":"ContainerStarted","Data":"ab16275379298f07255f08f892f3a6b560606b2415c65dca25931b72d1e2725f"} Nov 27 17:42:45 crc kubenswrapper[4809]: I1127 17:42:45.390273 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" podStartSLOduration=2.390246402 podStartE2EDuration="2.390246402s" podCreationTimestamp="2025-11-27 17:42:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:42:45.383829957 +0000 UTC m=+2000.656287319" watchObservedRunningTime="2025-11-27 17:42:45.390246402 +0000 UTC m=+2000.662703754" Nov 27 17:42:45 crc kubenswrapper[4809]: I1127 17:42:45.469331 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a8fdea-bcf5-49a2-8b88-13875ac05c71" path="/var/lib/kubelet/pods/14a8fdea-bcf5-49a2-8b88-13875ac05c71/volumes" Nov 27 17:42:53 crc kubenswrapper[4809]: I1127 17:42:53.642005 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-zgq88" Nov 27 17:42:53 crc kubenswrapper[4809]: I1127 17:42:53.711859 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-wwxf2"] Nov 27 17:42:53 crc kubenswrapper[4809]: I1127 17:42:53.712091 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" podUID="999823be-7414-421c-8ec1-dfeb6981546a" containerName="dnsmasq-dns" containerID="cri-o://89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a" gracePeriod=10 Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.203375 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.342694 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-config\") pod \"999823be-7414-421c-8ec1-dfeb6981546a\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.342809 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-swift-storage-0\") pod \"999823be-7414-421c-8ec1-dfeb6981546a\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.342843 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-svc\") pod \"999823be-7414-421c-8ec1-dfeb6981546a\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.342907 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-sb\") pod \"999823be-7414-421c-8ec1-dfeb6981546a\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.343619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-nb\") pod \"999823be-7414-421c-8ec1-dfeb6981546a\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.343903 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-openstack-edpm-ipam\") pod \"999823be-7414-421c-8ec1-dfeb6981546a\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.344162 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4bqv\" (UniqueName: \"kubernetes.io/projected/999823be-7414-421c-8ec1-dfeb6981546a-kube-api-access-x4bqv\") pod \"999823be-7414-421c-8ec1-dfeb6981546a\" (UID: \"999823be-7414-421c-8ec1-dfeb6981546a\") " Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.351082 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/999823be-7414-421c-8ec1-dfeb6981546a-kube-api-access-x4bqv" (OuterVolumeSpecName: "kube-api-access-x4bqv") pod "999823be-7414-421c-8ec1-dfeb6981546a" (UID: "999823be-7414-421c-8ec1-dfeb6981546a"). InnerVolumeSpecName "kube-api-access-x4bqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.411872 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "999823be-7414-421c-8ec1-dfeb6981546a" (UID: "999823be-7414-421c-8ec1-dfeb6981546a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.414258 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "999823be-7414-421c-8ec1-dfeb6981546a" (UID: "999823be-7414-421c-8ec1-dfeb6981546a"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.417940 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "999823be-7414-421c-8ec1-dfeb6981546a" (UID: "999823be-7414-421c-8ec1-dfeb6981546a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.421774 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "999823be-7414-421c-8ec1-dfeb6981546a" (UID: "999823be-7414-421c-8ec1-dfeb6981546a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.423161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "999823be-7414-421c-8ec1-dfeb6981546a" (UID: "999823be-7414-421c-8ec1-dfeb6981546a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.428674 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-config" (OuterVolumeSpecName: "config") pod "999823be-7414-421c-8ec1-dfeb6981546a" (UID: "999823be-7414-421c-8ec1-dfeb6981546a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.446558 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.446571 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" event={"ID":"999823be-7414-421c-8ec1-dfeb6981546a","Type":"ContainerDied","Data":"89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a"} Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.446672 4809 scope.go:117] "RemoveContainer" containerID="89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.446458 4809 generic.go:334] "Generic (PLEG): container finished" podID="999823be-7414-421c-8ec1-dfeb6981546a" containerID="89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a" exitCode=0 Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.447084 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-wwxf2" event={"ID":"999823be-7414-421c-8ec1-dfeb6981546a","Type":"ContainerDied","Data":"376f07d6830140f0e839ee02d2f16ba3e792d81267bb9f40ab8670eca6cb0f22"} Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.448097 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.449921 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4bqv\" (UniqueName: \"kubernetes.io/projected/999823be-7414-421c-8ec1-dfeb6981546a-kube-api-access-x4bqv\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.449956 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-config\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.449969 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.449982 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.450016 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.450028 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/999823be-7414-421c-8ec1-dfeb6981546a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.497759 4809 scope.go:117] "RemoveContainer" containerID="973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.499461 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-wwxf2"] Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.527126 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-wwxf2"] Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.532991 4809 scope.go:117] "RemoveContainer" containerID="89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a" Nov 27 17:42:54 crc kubenswrapper[4809]: E1127 17:42:54.533367 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a\": container with ID starting with 89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a not found: ID does not exist" containerID="89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.533410 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a"} err="failed to get container status \"89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a\": rpc error: code = NotFound desc = could not find container \"89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a\": container with ID starting with 89d32d74f96e187229cd08e0562356ad415d48cb24d5f2a05b8b1fd156c96a3a not found: ID does not exist" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.533436 4809 scope.go:117] "RemoveContainer" containerID="973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac" Nov 27 17:42:54 crc kubenswrapper[4809]: E1127 17:42:54.533841 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac\": container with ID starting with 973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac not found: ID does not exist" containerID="973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac" Nov 27 17:42:54 crc kubenswrapper[4809]: I1127 17:42:54.533877 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac"} err="failed to get container status \"973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac\": rpc error: code = NotFound desc = could not find container \"973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac\": container with ID starting with 973c928354bb48b4bdf0f665a05610abeccda3c86c148ef4fe658633e357b2ac not found: ID does not exist" Nov 27 17:42:55 crc kubenswrapper[4809]: I1127 17:42:55.469212 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="999823be-7414-421c-8ec1-dfeb6981546a" path="/var/lib/kubelet/pods/999823be-7414-421c-8ec1-dfeb6981546a/volumes" Nov 27 17:43:03 crc kubenswrapper[4809]: I1127 17:43:03.527686 4809 generic.go:334] "Generic (PLEG): container finished" podID="62edd287-129b-48b0-af61-d21bd27590d7" containerID="0428b26a5044722c21b49bf8b09b9c999ee5edac77f0e5e39491eeea99adee29" exitCode=0 Nov 27 17:43:03 crc kubenswrapper[4809]: I1127 17:43:03.527805 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62edd287-129b-48b0-af61-d21bd27590d7","Type":"ContainerDied","Data":"0428b26a5044722c21b49bf8b09b9c999ee5edac77f0e5e39491eeea99adee29"} Nov 27 17:43:04 crc kubenswrapper[4809]: I1127 17:43:04.539997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62edd287-129b-48b0-af61-d21bd27590d7","Type":"ContainerStarted","Data":"83626cc4ed834f3510c37fbdc330f5b0e1e38f7363504218046782ddf8952d5f"} Nov 27 17:43:04 crc kubenswrapper[4809]: I1127 17:43:04.540591 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 27 17:43:04 crc kubenswrapper[4809]: I1127 17:43:04.569046 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.569028348 podStartE2EDuration="36.569028348s" podCreationTimestamp="2025-11-27 17:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:43:04.567472365 +0000 UTC m=+2019.839929717" watchObservedRunningTime="2025-11-27 17:43:04.569028348 +0000 UTC m=+2019.841485700" Nov 27 17:43:05 crc kubenswrapper[4809]: I1127 17:43:05.551573 4809 generic.go:334] "Generic (PLEG): container finished" podID="3f0cad7c-85ca-4269-be47-555613716799" containerID="adba087f0b808b4f6fab80a9be3fa2d5a09a50e4f0b0614c8d64ab9a4fcac910" exitCode=0 Nov 27 17:43:05 crc kubenswrapper[4809]: I1127 17:43:05.551656 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3f0cad7c-85ca-4269-be47-555613716799","Type":"ContainerDied","Data":"adba087f0b808b4f6fab80a9be3fa2d5a09a50e4f0b0614c8d64ab9a4fcac910"} Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.563892 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3f0cad7c-85ca-4269-be47-555613716799","Type":"ContainerStarted","Data":"a402a67043ce3cb9db1e45dd82cb7f99ed8066d0b1c9859cf2eac5db9c8bb312"} Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.564414 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.588640 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.588618554 podStartE2EDuration="36.588618554s" podCreationTimestamp="2025-11-27 17:42:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 17:43:06.587533724 +0000 UTC m=+2021.859991086" watchObservedRunningTime="2025-11-27 17:43:06.588618554 +0000 UTC m=+2021.861075906" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.807827 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm"] Nov 27 17:43:06 crc kubenswrapper[4809]: E1127 17:43:06.808411 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a8fdea-bcf5-49a2-8b88-13875ac05c71" containerName="init" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.808435 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a8fdea-bcf5-49a2-8b88-13875ac05c71" containerName="init" Nov 27 17:43:06 crc kubenswrapper[4809]: E1127 17:43:06.808453 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="999823be-7414-421c-8ec1-dfeb6981546a" containerName="init" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.808461 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="999823be-7414-421c-8ec1-dfeb6981546a" containerName="init" Nov 27 17:43:06 crc kubenswrapper[4809]: E1127 17:43:06.808520 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a8fdea-bcf5-49a2-8b88-13875ac05c71" containerName="dnsmasq-dns" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.808530 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a8fdea-bcf5-49a2-8b88-13875ac05c71" containerName="dnsmasq-dns" Nov 27 17:43:06 crc kubenswrapper[4809]: E1127 17:43:06.808558 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="999823be-7414-421c-8ec1-dfeb6981546a" containerName="dnsmasq-dns" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.808568 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="999823be-7414-421c-8ec1-dfeb6981546a" containerName="dnsmasq-dns" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.808830 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a8fdea-bcf5-49a2-8b88-13875ac05c71" containerName="dnsmasq-dns" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.808874 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="999823be-7414-421c-8ec1-dfeb6981546a" containerName="dnsmasq-dns" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.809710 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.816438 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm"] Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.816687 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.821292 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.821300 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.821653 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.899074 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.899329 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdgjk\" (UniqueName: \"kubernetes.io/projected/56b0a916-b733-4a1c-b6dc-97925e319b83-kube-api-access-zdgjk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.899554 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:06 crc kubenswrapper[4809]: I1127 17:43:06.899669 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.001827 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.001911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.001974 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.002069 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdgjk\" (UniqueName: \"kubernetes.io/projected/56b0a916-b733-4a1c-b6dc-97925e319b83-kube-api-access-zdgjk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.006368 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.006848 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.006993 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.019382 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdgjk\" (UniqueName: \"kubernetes.io/projected/56b0a916-b733-4a1c-b6dc-97925e319b83-kube-api-access-zdgjk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.134103 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.654195 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm"] Nov 27 17:43:07 crc kubenswrapper[4809]: I1127 17:43:07.656846 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 17:43:08 crc kubenswrapper[4809]: I1127 17:43:08.584715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" event={"ID":"56b0a916-b733-4a1c-b6dc-97925e319b83","Type":"ContainerStarted","Data":"f33d8f4fe695e95d6abd36fd9b38f18e99a006aab2ead5c06023a23b7f04c367"} Nov 27 17:43:18 crc kubenswrapper[4809]: I1127 17:43:18.921944 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 27 17:43:19 crc kubenswrapper[4809]: I1127 17:43:19.719671 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" event={"ID":"56b0a916-b733-4a1c-b6dc-97925e319b83","Type":"ContainerStarted","Data":"51f07cc54dfe5ec131bf076cffc5f8dc8d9304863614401f9a8734f6130d287c"} Nov 27 17:43:19 crc kubenswrapper[4809]: I1127 17:43:19.746490 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" podStartSLOduration=1.928663291 podStartE2EDuration="13.746472474s" podCreationTimestamp="2025-11-27 17:43:06 +0000 UTC" firstStartedPulling="2025-11-27 17:43:07.656588728 +0000 UTC m=+2022.929046080" lastFinishedPulling="2025-11-27 17:43:19.474397911 +0000 UTC m=+2034.746855263" observedRunningTime="2025-11-27 17:43:19.739050553 +0000 UTC m=+2035.011507905" watchObservedRunningTime="2025-11-27 17:43:19.746472474 +0000 UTC m=+2035.018929826" Nov 27 17:43:21 crc kubenswrapper[4809]: I1127 17:43:21.184970 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 27 17:43:31 crc kubenswrapper[4809]: I1127 17:43:31.838482 4809 generic.go:334] "Generic (PLEG): container finished" podID="56b0a916-b733-4a1c-b6dc-97925e319b83" containerID="51f07cc54dfe5ec131bf076cffc5f8dc8d9304863614401f9a8734f6130d287c" exitCode=0 Nov 27 17:43:31 crc kubenswrapper[4809]: I1127 17:43:31.838582 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" event={"ID":"56b0a916-b733-4a1c-b6dc-97925e319b83","Type":"ContainerDied","Data":"51f07cc54dfe5ec131bf076cffc5f8dc8d9304863614401f9a8734f6130d287c"} Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.306243 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.331577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdgjk\" (UniqueName: \"kubernetes.io/projected/56b0a916-b733-4a1c-b6dc-97925e319b83-kube-api-access-zdgjk\") pod \"56b0a916-b733-4a1c-b6dc-97925e319b83\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.331818 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-repo-setup-combined-ca-bundle\") pod \"56b0a916-b733-4a1c-b6dc-97925e319b83\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.331953 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-inventory\") pod \"56b0a916-b733-4a1c-b6dc-97925e319b83\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.332107 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-ssh-key\") pod \"56b0a916-b733-4a1c-b6dc-97925e319b83\" (UID: \"56b0a916-b733-4a1c-b6dc-97925e319b83\") " Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.343305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "56b0a916-b733-4a1c-b6dc-97925e319b83" (UID: "56b0a916-b733-4a1c-b6dc-97925e319b83"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.343615 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56b0a916-b733-4a1c-b6dc-97925e319b83-kube-api-access-zdgjk" (OuterVolumeSpecName: "kube-api-access-zdgjk") pod "56b0a916-b733-4a1c-b6dc-97925e319b83" (UID: "56b0a916-b733-4a1c-b6dc-97925e319b83"). InnerVolumeSpecName "kube-api-access-zdgjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.365386 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "56b0a916-b733-4a1c-b6dc-97925e319b83" (UID: "56b0a916-b733-4a1c-b6dc-97925e319b83"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.377272 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-inventory" (OuterVolumeSpecName: "inventory") pod "56b0a916-b733-4a1c-b6dc-97925e319b83" (UID: "56b0a916-b733-4a1c-b6dc-97925e319b83"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.434398 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdgjk\" (UniqueName: \"kubernetes.io/projected/56b0a916-b733-4a1c-b6dc-97925e319b83-kube-api-access-zdgjk\") on node \"crc\" DevicePath \"\"" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.434431 4809 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.434442 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.434459 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b0a916-b733-4a1c-b6dc-97925e319b83-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.859791 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" event={"ID":"56b0a916-b733-4a1c-b6dc-97925e319b83","Type":"ContainerDied","Data":"f33d8f4fe695e95d6abd36fd9b38f18e99a006aab2ead5c06023a23b7f04c367"} Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.859839 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f33d8f4fe695e95d6abd36fd9b38f18e99a006aab2ead5c06023a23b7f04c367" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.859896 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.935786 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv"] Nov 27 17:43:33 crc kubenswrapper[4809]: E1127 17:43:33.936233 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56b0a916-b733-4a1c-b6dc-97925e319b83" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.936250 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="56b0a916-b733-4a1c-b6dc-97925e319b83" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.936486 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="56b0a916-b733-4a1c-b6dc-97925e319b83" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.937181 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.939265 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.939455 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.939505 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.941416 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:43:33 crc kubenswrapper[4809]: I1127 17:43:33.958681 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv"] Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.047461 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w5fcv\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.047534 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfchh\" (UniqueName: \"kubernetes.io/projected/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-kube-api-access-rfchh\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w5fcv\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.047773 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w5fcv\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.150501 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w5fcv\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.150662 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w5fcv\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.150723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfchh\" (UniqueName: \"kubernetes.io/projected/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-kube-api-access-rfchh\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w5fcv\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.156648 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w5fcv\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.156689 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w5fcv\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.171579 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfchh\" (UniqueName: \"kubernetes.io/projected/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-kube-api-access-rfchh\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w5fcv\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.256854 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:34 crc kubenswrapper[4809]: I1127 17:43:34.891476 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv"] Nov 27 17:43:35 crc kubenswrapper[4809]: I1127 17:43:35.879945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" event={"ID":"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e","Type":"ContainerStarted","Data":"f40cece3b2090bb362aeffb727ea2d87ffd38875d44cbb4a5cb81f5092843c3c"} Nov 27 17:43:36 crc kubenswrapper[4809]: I1127 17:43:36.889622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" event={"ID":"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e","Type":"ContainerStarted","Data":"20e0d0a1f64945c899cf601d2720a7098a90f065539e61979371c9b47e6b9f6e"} Nov 27 17:43:36 crc kubenswrapper[4809]: I1127 17:43:36.908490 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" podStartSLOduration=3.061013541 podStartE2EDuration="3.908471719s" podCreationTimestamp="2025-11-27 17:43:33 +0000 UTC" firstStartedPulling="2025-11-27 17:43:34.902354508 +0000 UTC m=+2050.174811860" lastFinishedPulling="2025-11-27 17:43:35.749812686 +0000 UTC m=+2051.022270038" observedRunningTime="2025-11-27 17:43:36.905768986 +0000 UTC m=+2052.178226338" watchObservedRunningTime="2025-11-27 17:43:36.908471719 +0000 UTC m=+2052.180929071" Nov 27 17:43:38 crc kubenswrapper[4809]: I1127 17:43:38.911753 4809 generic.go:334] "Generic (PLEG): container finished" podID="cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e" containerID="20e0d0a1f64945c899cf601d2720a7098a90f065539e61979371c9b47e6b9f6e" exitCode=0 Nov 27 17:43:38 crc kubenswrapper[4809]: I1127 17:43:38.911816 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" event={"ID":"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e","Type":"ContainerDied","Data":"20e0d0a1f64945c899cf601d2720a7098a90f065539e61979371c9b47e6b9f6e"} Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.356876 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.388549 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-inventory\") pod \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.388628 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-ssh-key\") pod \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.388670 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfchh\" (UniqueName: \"kubernetes.io/projected/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-kube-api-access-rfchh\") pod \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\" (UID: \"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e\") " Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.395334 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-kube-api-access-rfchh" (OuterVolumeSpecName: "kube-api-access-rfchh") pod "cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e" (UID: "cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e"). InnerVolumeSpecName "kube-api-access-rfchh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.416969 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-inventory" (OuterVolumeSpecName: "inventory") pod "cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e" (UID: "cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.419748 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e" (UID: "cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.492387 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.492600 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.492702 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfchh\" (UniqueName: \"kubernetes.io/projected/cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e-kube-api-access-rfchh\") on node \"crc\" DevicePath \"\"" Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.933809 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" event={"ID":"cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e","Type":"ContainerDied","Data":"f40cece3b2090bb362aeffb727ea2d87ffd38875d44cbb4a5cb81f5092843c3c"} Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.934102 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f40cece3b2090bb362aeffb727ea2d87ffd38875d44cbb4a5cb81f5092843c3c" Nov 27 17:43:40 crc kubenswrapper[4809]: I1127 17:43:40.933886 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w5fcv" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.041464 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq"] Nov 27 17:43:41 crc kubenswrapper[4809]: E1127 17:43:41.041967 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.041989 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.042269 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.043164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.045254 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.045254 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.046021 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.046653 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.052981 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq"] Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.112003 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p65sh\" (UniqueName: \"kubernetes.io/projected/1f21abe6-3443-4a72-b93b-574203c3f3b4-kube-api-access-p65sh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.112104 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.112310 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.112335 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.214876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.215263 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.215435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p65sh\" (UniqueName: \"kubernetes.io/projected/1f21abe6-3443-4a72-b93b-574203c3f3b4-kube-api-access-p65sh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.215520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.220769 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.221167 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.233348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.241515 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p65sh\" (UniqueName: \"kubernetes.io/projected/1f21abe6-3443-4a72-b93b-574203c3f3b4-kube-api-access-p65sh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.361806 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.909193 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq"] Nov 27 17:43:41 crc kubenswrapper[4809]: I1127 17:43:41.943636 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" event={"ID":"1f21abe6-3443-4a72-b93b-574203c3f3b4","Type":"ContainerStarted","Data":"3ed817c6e2f16423730660b37318e89b13fae709d50a6e24582eb01e803f47d7"} Nov 27 17:43:42 crc kubenswrapper[4809]: I1127 17:43:42.953931 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" event={"ID":"1f21abe6-3443-4a72-b93b-574203c3f3b4","Type":"ContainerStarted","Data":"75c9f6684299834884de5cfec27c8274a764885f0ae53187932e0f81db67437d"} Nov 27 17:43:42 crc kubenswrapper[4809]: I1127 17:43:42.970082 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" podStartSLOduration=1.582717235 podStartE2EDuration="1.970064843s" podCreationTimestamp="2025-11-27 17:43:41 +0000 UTC" firstStartedPulling="2025-11-27 17:43:41.910472276 +0000 UTC m=+2057.182929618" lastFinishedPulling="2025-11-27 17:43:42.297819874 +0000 UTC m=+2057.570277226" observedRunningTime="2025-11-27 17:43:42.96847564 +0000 UTC m=+2058.240932992" watchObservedRunningTime="2025-11-27 17:43:42.970064843 +0000 UTC m=+2058.242522195" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.002635 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-svlt4"] Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.005409 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.021386 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svlt4"] Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.089436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-utilities\") pod \"community-operators-svlt4\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.089799 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-catalog-content\") pod \"community-operators-svlt4\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.089873 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l6fb\" (UniqueName: \"kubernetes.io/projected/53132ebd-25bb-4024-98d2-ea9a9821df3b-kube-api-access-5l6fb\") pod \"community-operators-svlt4\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.191966 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-utilities\") pod \"community-operators-svlt4\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.192433 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-utilities\") pod \"community-operators-svlt4\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.192570 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-catalog-content\") pod \"community-operators-svlt4\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.192660 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l6fb\" (UniqueName: \"kubernetes.io/projected/53132ebd-25bb-4024-98d2-ea9a9821df3b-kube-api-access-5l6fb\") pod \"community-operators-svlt4\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.192913 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-catalog-content\") pod \"community-operators-svlt4\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.213881 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l6fb\" (UniqueName: \"kubernetes.io/projected/53132ebd-25bb-4024-98d2-ea9a9821df3b-kube-api-access-5l6fb\") pod \"community-operators-svlt4\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.330313 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:00 crc kubenswrapper[4809]: I1127 17:44:00.923955 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svlt4"] Nov 27 17:44:01 crc kubenswrapper[4809]: I1127 17:44:01.122150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlt4" event={"ID":"53132ebd-25bb-4024-98d2-ea9a9821df3b","Type":"ContainerStarted","Data":"6dbf68001f0a2379e9fabf6cc2864ef60450250451fd3a8257d897be01ac835f"} Nov 27 17:44:02 crc kubenswrapper[4809]: I1127 17:44:02.133498 4809 generic.go:334] "Generic (PLEG): container finished" podID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerID="b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8" exitCode=0 Nov 27 17:44:02 crc kubenswrapper[4809]: I1127 17:44:02.133580 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlt4" event={"ID":"53132ebd-25bb-4024-98d2-ea9a9821df3b","Type":"ContainerDied","Data":"b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8"} Nov 27 17:44:03 crc kubenswrapper[4809]: I1127 17:44:03.146563 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlt4" event={"ID":"53132ebd-25bb-4024-98d2-ea9a9821df3b","Type":"ContainerStarted","Data":"cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a"} Nov 27 17:44:04 crc kubenswrapper[4809]: I1127 17:44:04.155685 4809 generic.go:334] "Generic (PLEG): container finished" podID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerID="cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a" exitCode=0 Nov 27 17:44:04 crc kubenswrapper[4809]: I1127 17:44:04.155768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlt4" event={"ID":"53132ebd-25bb-4024-98d2-ea9a9821df3b","Type":"ContainerDied","Data":"cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a"} Nov 27 17:44:05 crc kubenswrapper[4809]: I1127 17:44:05.166381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlt4" event={"ID":"53132ebd-25bb-4024-98d2-ea9a9821df3b","Type":"ContainerStarted","Data":"b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190"} Nov 27 17:44:05 crc kubenswrapper[4809]: I1127 17:44:05.191669 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-svlt4" podStartSLOduration=3.60617752 podStartE2EDuration="6.191648215s" podCreationTimestamp="2025-11-27 17:43:59 +0000 UTC" firstStartedPulling="2025-11-27 17:44:02.135528895 +0000 UTC m=+2077.407986247" lastFinishedPulling="2025-11-27 17:44:04.72099959 +0000 UTC m=+2079.993456942" observedRunningTime="2025-11-27 17:44:05.182032995 +0000 UTC m=+2080.454490347" watchObservedRunningTime="2025-11-27 17:44:05.191648215 +0000 UTC m=+2080.464105567" Nov 27 17:44:10 crc kubenswrapper[4809]: I1127 17:44:10.330640 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:10 crc kubenswrapper[4809]: I1127 17:44:10.331241 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:10 crc kubenswrapper[4809]: I1127 17:44:10.376635 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:11 crc kubenswrapper[4809]: I1127 17:44:11.277099 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:11 crc kubenswrapper[4809]: I1127 17:44:11.382923 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svlt4"] Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.246333 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-svlt4" podUID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerName="registry-server" containerID="cri-o://b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190" gracePeriod=2 Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.715352 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.874283 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-catalog-content\") pod \"53132ebd-25bb-4024-98d2-ea9a9821df3b\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.874359 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l6fb\" (UniqueName: \"kubernetes.io/projected/53132ebd-25bb-4024-98d2-ea9a9821df3b-kube-api-access-5l6fb\") pod \"53132ebd-25bb-4024-98d2-ea9a9821df3b\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.874402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-utilities\") pod \"53132ebd-25bb-4024-98d2-ea9a9821df3b\" (UID: \"53132ebd-25bb-4024-98d2-ea9a9821df3b\") " Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.875170 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-utilities" (OuterVolumeSpecName: "utilities") pod "53132ebd-25bb-4024-98d2-ea9a9821df3b" (UID: "53132ebd-25bb-4024-98d2-ea9a9821df3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.879580 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53132ebd-25bb-4024-98d2-ea9a9821df3b-kube-api-access-5l6fb" (OuterVolumeSpecName: "kube-api-access-5l6fb") pod "53132ebd-25bb-4024-98d2-ea9a9821df3b" (UID: "53132ebd-25bb-4024-98d2-ea9a9821df3b"). InnerVolumeSpecName "kube-api-access-5l6fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.920353 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53132ebd-25bb-4024-98d2-ea9a9821df3b" (UID: "53132ebd-25bb-4024-98d2-ea9a9821df3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.977166 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.977195 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ebd-25bb-4024-98d2-ea9a9821df3b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:44:13 crc kubenswrapper[4809]: I1127 17:44:13.977208 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l6fb\" (UniqueName: \"kubernetes.io/projected/53132ebd-25bb-4024-98d2-ea9a9821df3b-kube-api-access-5l6fb\") on node \"crc\" DevicePath \"\"" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.257847 4809 generic.go:334] "Generic (PLEG): container finished" podID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerID="b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190" exitCode=0 Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.257894 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svlt4" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.257907 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlt4" event={"ID":"53132ebd-25bb-4024-98d2-ea9a9821df3b","Type":"ContainerDied","Data":"b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190"} Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.258730 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlt4" event={"ID":"53132ebd-25bb-4024-98d2-ea9a9821df3b","Type":"ContainerDied","Data":"6dbf68001f0a2379e9fabf6cc2864ef60450250451fd3a8257d897be01ac835f"} Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.258782 4809 scope.go:117] "RemoveContainer" containerID="b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.279928 4809 scope.go:117] "RemoveContainer" containerID="cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.290545 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svlt4"] Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.299303 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-svlt4"] Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.313735 4809 scope.go:117] "RemoveContainer" containerID="b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.352823 4809 scope.go:117] "RemoveContainer" containerID="b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190" Nov 27 17:44:14 crc kubenswrapper[4809]: E1127 17:44:14.353275 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190\": container with ID starting with b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190 not found: ID does not exist" containerID="b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.353316 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190"} err="failed to get container status \"b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190\": rpc error: code = NotFound desc = could not find container \"b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190\": container with ID starting with b0a40aedf76593fcffa07a35428f9eda7df0eb3a48d1c959c6bb771aca129190 not found: ID does not exist" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.353342 4809 scope.go:117] "RemoveContainer" containerID="cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a" Nov 27 17:44:14 crc kubenswrapper[4809]: E1127 17:44:14.353751 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a\": container with ID starting with cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a not found: ID does not exist" containerID="cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.353775 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a"} err="failed to get container status \"cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a\": rpc error: code = NotFound desc = could not find container \"cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a\": container with ID starting with cc6015030e78647382043084ede25c5198e594c9e0018ef3b5389f95e500e54a not found: ID does not exist" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.353790 4809 scope.go:117] "RemoveContainer" containerID="b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8" Nov 27 17:44:14 crc kubenswrapper[4809]: E1127 17:44:14.354118 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8\": container with ID starting with b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8 not found: ID does not exist" containerID="b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8" Nov 27 17:44:14 crc kubenswrapper[4809]: I1127 17:44:14.354219 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8"} err="failed to get container status \"b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8\": rpc error: code = NotFound desc = could not find container \"b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8\": container with ID starting with b3fa3f106de77bee6c53fe8bc4f8ab85a0e3ed741d1717d1f783832014f96be8 not found: ID does not exist" Nov 27 17:44:15 crc kubenswrapper[4809]: I1127 17:44:15.494370 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53132ebd-25bb-4024-98d2-ea9a9821df3b" path="/var/lib/kubelet/pods/53132ebd-25bb-4024-98d2-ea9a9821df3b/volumes" Nov 27 17:44:32 crc kubenswrapper[4809]: I1127 17:44:32.644832 4809 scope.go:117] "RemoveContainer" containerID="641fe1a719780307288b3c357021f5dd62ba4d1d0e1af01d4cef3bb1bf38dff5" Nov 27 17:44:55 crc kubenswrapper[4809]: I1127 17:44:55.779874 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:44:55 crc kubenswrapper[4809]: I1127 17:44:55.780427 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.151244 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d"] Nov 27 17:45:00 crc kubenswrapper[4809]: E1127 17:45:00.152237 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerName="extract-content" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.152250 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerName="extract-content" Nov 27 17:45:00 crc kubenswrapper[4809]: E1127 17:45:00.152262 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerName="extract-utilities" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.152271 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerName="extract-utilities" Nov 27 17:45:00 crc kubenswrapper[4809]: E1127 17:45:00.152281 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerName="registry-server" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.152287 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerName="registry-server" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.152492 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="53132ebd-25bb-4024-98d2-ea9a9821df3b" containerName="registry-server" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.153296 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.156171 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.156226 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.170636 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d"] Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.261084 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a5428a8-1555-48ba-9b96-836b994671c0-config-volume\") pod \"collect-profiles-29404425-j5p8d\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.261515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a5428a8-1555-48ba-9b96-836b994671c0-secret-volume\") pod \"collect-profiles-29404425-j5p8d\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.261642 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjr45\" (UniqueName: \"kubernetes.io/projected/8a5428a8-1555-48ba-9b96-836b994671c0-kube-api-access-cjr45\") pod \"collect-profiles-29404425-j5p8d\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.363960 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a5428a8-1555-48ba-9b96-836b994671c0-secret-volume\") pod \"collect-profiles-29404425-j5p8d\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.364012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjr45\" (UniqueName: \"kubernetes.io/projected/8a5428a8-1555-48ba-9b96-836b994671c0-kube-api-access-cjr45\") pod \"collect-profiles-29404425-j5p8d\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.364058 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a5428a8-1555-48ba-9b96-836b994671c0-config-volume\") pod \"collect-profiles-29404425-j5p8d\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.365174 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a5428a8-1555-48ba-9b96-836b994671c0-config-volume\") pod \"collect-profiles-29404425-j5p8d\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.370920 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a5428a8-1555-48ba-9b96-836b994671c0-secret-volume\") pod \"collect-profiles-29404425-j5p8d\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.381751 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjr45\" (UniqueName: \"kubernetes.io/projected/8a5428a8-1555-48ba-9b96-836b994671c0-kube-api-access-cjr45\") pod \"collect-profiles-29404425-j5p8d\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.479235 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:00 crc kubenswrapper[4809]: I1127 17:45:00.921291 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d"] Nov 27 17:45:01 crc kubenswrapper[4809]: I1127 17:45:01.692270 4809 generic.go:334] "Generic (PLEG): container finished" podID="8a5428a8-1555-48ba-9b96-836b994671c0" containerID="ab9798b43ad5ac4217559d85dfa0a52119872ddca0399d941260284113de49f8" exitCode=0 Nov 27 17:45:01 crc kubenswrapper[4809]: I1127 17:45:01.692369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" event={"ID":"8a5428a8-1555-48ba-9b96-836b994671c0","Type":"ContainerDied","Data":"ab9798b43ad5ac4217559d85dfa0a52119872ddca0399d941260284113de49f8"} Nov 27 17:45:01 crc kubenswrapper[4809]: I1127 17:45:01.692636 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" event={"ID":"8a5428a8-1555-48ba-9b96-836b994671c0","Type":"ContainerStarted","Data":"1e9319babb6356abed2f11893dfc3158cc13797120c59939cadbef3d3613c5ae"} Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.094626 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.125926 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjr45\" (UniqueName: \"kubernetes.io/projected/8a5428a8-1555-48ba-9b96-836b994671c0-kube-api-access-cjr45\") pod \"8a5428a8-1555-48ba-9b96-836b994671c0\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.125973 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a5428a8-1555-48ba-9b96-836b994671c0-config-volume\") pod \"8a5428a8-1555-48ba-9b96-836b994671c0\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.126296 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a5428a8-1555-48ba-9b96-836b994671c0-secret-volume\") pod \"8a5428a8-1555-48ba-9b96-836b994671c0\" (UID: \"8a5428a8-1555-48ba-9b96-836b994671c0\") " Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.126857 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5428a8-1555-48ba-9b96-836b994671c0-config-volume" (OuterVolumeSpecName: "config-volume") pod "8a5428a8-1555-48ba-9b96-836b994671c0" (UID: "8a5428a8-1555-48ba-9b96-836b994671c0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.131962 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5428a8-1555-48ba-9b96-836b994671c0-kube-api-access-cjr45" (OuterVolumeSpecName: "kube-api-access-cjr45") pod "8a5428a8-1555-48ba-9b96-836b994671c0" (UID: "8a5428a8-1555-48ba-9b96-836b994671c0"). InnerVolumeSpecName "kube-api-access-cjr45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.138158 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5428a8-1555-48ba-9b96-836b994671c0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8a5428a8-1555-48ba-9b96-836b994671c0" (UID: "8a5428a8-1555-48ba-9b96-836b994671c0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.228188 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a5428a8-1555-48ba-9b96-836b994671c0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.228234 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjr45\" (UniqueName: \"kubernetes.io/projected/8a5428a8-1555-48ba-9b96-836b994671c0-kube-api-access-cjr45\") on node \"crc\" DevicePath \"\"" Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.228248 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a5428a8-1555-48ba-9b96-836b994671c0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.712611 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" event={"ID":"8a5428a8-1555-48ba-9b96-836b994671c0","Type":"ContainerDied","Data":"1e9319babb6356abed2f11893dfc3158cc13797120c59939cadbef3d3613c5ae"} Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.712666 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e9319babb6356abed2f11893dfc3158cc13797120c59939cadbef3d3613c5ae" Nov 27 17:45:03 crc kubenswrapper[4809]: I1127 17:45:03.712680 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d" Nov 27 17:45:04 crc kubenswrapper[4809]: I1127 17:45:04.170811 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd"] Nov 27 17:45:04 crc kubenswrapper[4809]: I1127 17:45:04.183515 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404380-j6wjd"] Nov 27 17:45:05 crc kubenswrapper[4809]: I1127 17:45:05.470533 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15c3167d-77d3-457f-b602-5e87f96afe8d" path="/var/lib/kubelet/pods/15c3167d-77d3-457f-b602-5e87f96afe8d/volumes" Nov 27 17:45:25 crc kubenswrapper[4809]: I1127 17:45:25.779104 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:45:25 crc kubenswrapper[4809]: I1127 17:45:25.779654 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:45:32 crc kubenswrapper[4809]: I1127 17:45:32.715926 4809 scope.go:117] "RemoveContainer" containerID="c15b4e7c3ade0d28dabdfa9a5c9de9de7e4eba3b3ffca27b1e5251c7acbbb3c7" Nov 27 17:45:32 crc kubenswrapper[4809]: I1127 17:45:32.746581 4809 scope.go:117] "RemoveContainer" containerID="c8fb2111f80e8c44f310c5346c2a3ee96da1beaf0221cf3591832a76006b7ad0" Nov 27 17:45:32 crc kubenswrapper[4809]: I1127 17:45:32.814479 4809 scope.go:117] "RemoveContainer" containerID="3c3714c1a6319450a3525d33ff2a1a82e6ca523ba356cf684a9d85b33b861836" Nov 27 17:45:55 crc kubenswrapper[4809]: I1127 17:45:55.779863 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:45:55 crc kubenswrapper[4809]: I1127 17:45:55.780485 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:45:55 crc kubenswrapper[4809]: I1127 17:45:55.780537 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:45:55 crc kubenswrapper[4809]: I1127 17:45:55.781202 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9a6ef73344f0f4b1da4284fd6ac88d1404b7c5e1410dce4a9042ec288f5d1365"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:45:55 crc kubenswrapper[4809]: I1127 17:45:55.781253 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://9a6ef73344f0f4b1da4284fd6ac88d1404b7c5e1410dce4a9042ec288f5d1365" gracePeriod=600 Nov 27 17:45:56 crc kubenswrapper[4809]: I1127 17:45:56.192132 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="9a6ef73344f0f4b1da4284fd6ac88d1404b7c5e1410dce4a9042ec288f5d1365" exitCode=0 Nov 27 17:45:56 crc kubenswrapper[4809]: I1127 17:45:56.192209 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"9a6ef73344f0f4b1da4284fd6ac88d1404b7c5e1410dce4a9042ec288f5d1365"} Nov 27 17:45:56 crc kubenswrapper[4809]: I1127 17:45:56.192519 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0"} Nov 27 17:45:56 crc kubenswrapper[4809]: I1127 17:45:56.192555 4809 scope.go:117] "RemoveContainer" containerID="43e56b69adc94f0828e2c211aad2f6e2d6c7af05601da712ca7198b3a7c87596" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.689554 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zrdrg"] Nov 27 17:46:30 crc kubenswrapper[4809]: E1127 17:46:30.690525 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5428a8-1555-48ba-9b96-836b994671c0" containerName="collect-profiles" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.690537 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5428a8-1555-48ba-9b96-836b994671c0" containerName="collect-profiles" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.690758 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5428a8-1555-48ba-9b96-836b994671c0" containerName="collect-profiles" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.692262 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.702249 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zrdrg"] Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.778878 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-catalog-content\") pod \"certified-operators-zrdrg\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.778946 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-utilities\") pod \"certified-operators-zrdrg\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.779057 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgh7b\" (UniqueName: \"kubernetes.io/projected/5a93d1ed-6bf0-46af-a351-fa067e36ad34-kube-api-access-xgh7b\") pod \"certified-operators-zrdrg\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.880961 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgh7b\" (UniqueName: \"kubernetes.io/projected/5a93d1ed-6bf0-46af-a351-fa067e36ad34-kube-api-access-xgh7b\") pod \"certified-operators-zrdrg\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.881058 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-catalog-content\") pod \"certified-operators-zrdrg\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.881102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-utilities\") pod \"certified-operators-zrdrg\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.881726 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-utilities\") pod \"certified-operators-zrdrg\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.882280 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-catalog-content\") pod \"certified-operators-zrdrg\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:30 crc kubenswrapper[4809]: I1127 17:46:30.902797 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgh7b\" (UniqueName: \"kubernetes.io/projected/5a93d1ed-6bf0-46af-a351-fa067e36ad34-kube-api-access-xgh7b\") pod \"certified-operators-zrdrg\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:31 crc kubenswrapper[4809]: I1127 17:46:31.013271 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:31 crc kubenswrapper[4809]: I1127 17:46:31.566759 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zrdrg"] Nov 27 17:46:32 crc kubenswrapper[4809]: I1127 17:46:32.564764 4809 generic.go:334] "Generic (PLEG): container finished" podID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerID="382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009" exitCode=0 Nov 27 17:46:32 crc kubenswrapper[4809]: I1127 17:46:32.564881 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrdrg" event={"ID":"5a93d1ed-6bf0-46af-a351-fa067e36ad34","Type":"ContainerDied","Data":"382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009"} Nov 27 17:46:32 crc kubenswrapper[4809]: I1127 17:46:32.565962 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrdrg" event={"ID":"5a93d1ed-6bf0-46af-a351-fa067e36ad34","Type":"ContainerStarted","Data":"f5106a24b0f360c050b0859293c3f638ada93ca756034bc521c8ad200325c7e9"} Nov 27 17:46:33 crc kubenswrapper[4809]: I1127 17:46:33.578288 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrdrg" event={"ID":"5a93d1ed-6bf0-46af-a351-fa067e36ad34","Type":"ContainerStarted","Data":"10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3"} Nov 27 17:46:34 crc kubenswrapper[4809]: I1127 17:46:34.601258 4809 generic.go:334] "Generic (PLEG): container finished" podID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerID="10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3" exitCode=0 Nov 27 17:46:34 crc kubenswrapper[4809]: I1127 17:46:34.601319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrdrg" event={"ID":"5a93d1ed-6bf0-46af-a351-fa067e36ad34","Type":"ContainerDied","Data":"10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3"} Nov 27 17:46:36 crc kubenswrapper[4809]: I1127 17:46:36.622127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrdrg" event={"ID":"5a93d1ed-6bf0-46af-a351-fa067e36ad34","Type":"ContainerStarted","Data":"2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558"} Nov 27 17:46:36 crc kubenswrapper[4809]: I1127 17:46:36.648105 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zrdrg" podStartSLOduration=3.782763496 podStartE2EDuration="6.648075713s" podCreationTimestamp="2025-11-27 17:46:30 +0000 UTC" firstStartedPulling="2025-11-27 17:46:32.567550592 +0000 UTC m=+2227.840007944" lastFinishedPulling="2025-11-27 17:46:35.432862809 +0000 UTC m=+2230.705320161" observedRunningTime="2025-11-27 17:46:36.638113491 +0000 UTC m=+2231.910570853" watchObservedRunningTime="2025-11-27 17:46:36.648075713 +0000 UTC m=+2231.920533075" Nov 27 17:46:41 crc kubenswrapper[4809]: I1127 17:46:41.013867 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:41 crc kubenswrapper[4809]: I1127 17:46:41.014454 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:41 crc kubenswrapper[4809]: I1127 17:46:41.061437 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:41 crc kubenswrapper[4809]: I1127 17:46:41.722402 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:41 crc kubenswrapper[4809]: I1127 17:46:41.785667 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zrdrg"] Nov 27 17:46:43 crc kubenswrapper[4809]: I1127 17:46:43.696401 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zrdrg" podUID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerName="registry-server" containerID="cri-o://2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558" gracePeriod=2 Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.170308 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.250155 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgh7b\" (UniqueName: \"kubernetes.io/projected/5a93d1ed-6bf0-46af-a351-fa067e36ad34-kube-api-access-xgh7b\") pod \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.250352 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-utilities\") pod \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.250380 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-catalog-content\") pod \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\" (UID: \"5a93d1ed-6bf0-46af-a351-fa067e36ad34\") " Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.252008 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-utilities" (OuterVolumeSpecName: "utilities") pod "5a93d1ed-6bf0-46af-a351-fa067e36ad34" (UID: "5a93d1ed-6bf0-46af-a351-fa067e36ad34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.256246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a93d1ed-6bf0-46af-a351-fa067e36ad34-kube-api-access-xgh7b" (OuterVolumeSpecName: "kube-api-access-xgh7b") pod "5a93d1ed-6bf0-46af-a351-fa067e36ad34" (UID: "5a93d1ed-6bf0-46af-a351-fa067e36ad34"). InnerVolumeSpecName "kube-api-access-xgh7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.295480 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a93d1ed-6bf0-46af-a351-fa067e36ad34" (UID: "5a93d1ed-6bf0-46af-a351-fa067e36ad34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.352596 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.352636 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a93d1ed-6bf0-46af-a351-fa067e36ad34-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.352648 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgh7b\" (UniqueName: \"kubernetes.io/projected/5a93d1ed-6bf0-46af-a351-fa067e36ad34-kube-api-access-xgh7b\") on node \"crc\" DevicePath \"\"" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.705689 4809 generic.go:334] "Generic (PLEG): container finished" podID="1f21abe6-3443-4a72-b93b-574203c3f3b4" containerID="75c9f6684299834884de5cfec27c8274a764885f0ae53187932e0f81db67437d" exitCode=0 Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.705775 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" event={"ID":"1f21abe6-3443-4a72-b93b-574203c3f3b4","Type":"ContainerDied","Data":"75c9f6684299834884de5cfec27c8274a764885f0ae53187932e0f81db67437d"} Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.708642 4809 generic.go:334] "Generic (PLEG): container finished" podID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerID="2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558" exitCode=0 Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.708674 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrdrg" event={"ID":"5a93d1ed-6bf0-46af-a351-fa067e36ad34","Type":"ContainerDied","Data":"2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558"} Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.708690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrdrg" event={"ID":"5a93d1ed-6bf0-46af-a351-fa067e36ad34","Type":"ContainerDied","Data":"f5106a24b0f360c050b0859293c3f638ada93ca756034bc521c8ad200325c7e9"} Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.708709 4809 scope.go:117] "RemoveContainer" containerID="2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.708721 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrdrg" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.739445 4809 scope.go:117] "RemoveContainer" containerID="10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.746794 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zrdrg"] Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.756540 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zrdrg"] Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.761033 4809 scope.go:117] "RemoveContainer" containerID="382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.810033 4809 scope.go:117] "RemoveContainer" containerID="2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558" Nov 27 17:46:44 crc kubenswrapper[4809]: E1127 17:46:44.810996 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558\": container with ID starting with 2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558 not found: ID does not exist" containerID="2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.811067 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558"} err="failed to get container status \"2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558\": rpc error: code = NotFound desc = could not find container \"2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558\": container with ID starting with 2a6f01e678ffc1f5fa0b41d6337dc336a56627f20c82e1daf203131e9de5c558 not found: ID does not exist" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.811118 4809 scope.go:117] "RemoveContainer" containerID="10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3" Nov 27 17:46:44 crc kubenswrapper[4809]: E1127 17:46:44.811691 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3\": container with ID starting with 10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3 not found: ID does not exist" containerID="10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.811730 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3"} err="failed to get container status \"10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3\": rpc error: code = NotFound desc = could not find container \"10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3\": container with ID starting with 10633bf6d736218ae4ba05437207b504a41fd346f3e17fd1cd14460384d4a3a3 not found: ID does not exist" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.811777 4809 scope.go:117] "RemoveContainer" containerID="382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009" Nov 27 17:46:44 crc kubenswrapper[4809]: E1127 17:46:44.812060 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009\": container with ID starting with 382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009 not found: ID does not exist" containerID="382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009" Nov 27 17:46:44 crc kubenswrapper[4809]: I1127 17:46:44.812093 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009"} err="failed to get container status \"382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009\": rpc error: code = NotFound desc = could not find container \"382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009\": container with ID starting with 382ea05c1fa402b650e980da6704cb3a1d75a66e96a8f9ebdcf26b1d9ec11009 not found: ID does not exist" Nov 27 17:46:45 crc kubenswrapper[4809]: I1127 17:46:45.469933 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" path="/var/lib/kubelet/pods/5a93d1ed-6bf0-46af-a351-fa067e36ad34/volumes" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.164682 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.289404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p65sh\" (UniqueName: \"kubernetes.io/projected/1f21abe6-3443-4a72-b93b-574203c3f3b4-kube-api-access-p65sh\") pod \"1f21abe6-3443-4a72-b93b-574203c3f3b4\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.289977 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-bootstrap-combined-ca-bundle\") pod \"1f21abe6-3443-4a72-b93b-574203c3f3b4\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.290207 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-ssh-key\") pod \"1f21abe6-3443-4a72-b93b-574203c3f3b4\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.290342 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-inventory\") pod \"1f21abe6-3443-4a72-b93b-574203c3f3b4\" (UID: \"1f21abe6-3443-4a72-b93b-574203c3f3b4\") " Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.296431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f21abe6-3443-4a72-b93b-574203c3f3b4-kube-api-access-p65sh" (OuterVolumeSpecName: "kube-api-access-p65sh") pod "1f21abe6-3443-4a72-b93b-574203c3f3b4" (UID: "1f21abe6-3443-4a72-b93b-574203c3f3b4"). InnerVolumeSpecName "kube-api-access-p65sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.296999 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1f21abe6-3443-4a72-b93b-574203c3f3b4" (UID: "1f21abe6-3443-4a72-b93b-574203c3f3b4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.322578 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-inventory" (OuterVolumeSpecName: "inventory") pod "1f21abe6-3443-4a72-b93b-574203c3f3b4" (UID: "1f21abe6-3443-4a72-b93b-574203c3f3b4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.333464 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1f21abe6-3443-4a72-b93b-574203c3f3b4" (UID: "1f21abe6-3443-4a72-b93b-574203c3f3b4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.393027 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.393059 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.393069 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p65sh\" (UniqueName: \"kubernetes.io/projected/1f21abe6-3443-4a72-b93b-574203c3f3b4-kube-api-access-p65sh\") on node \"crc\" DevicePath \"\"" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.393082 4809 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f21abe6-3443-4a72-b93b-574203c3f3b4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.738761 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" event={"ID":"1f21abe6-3443-4a72-b93b-574203c3f3b4","Type":"ContainerDied","Data":"3ed817c6e2f16423730660b37318e89b13fae709d50a6e24582eb01e803f47d7"} Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.738800 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ed817c6e2f16423730660b37318e89b13fae709d50a6e24582eb01e803f47d7" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.738841 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.799724 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj"] Nov 27 17:46:46 crc kubenswrapper[4809]: E1127 17:46:46.801606 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerName="extract-content" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.801626 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerName="extract-content" Nov 27 17:46:46 crc kubenswrapper[4809]: E1127 17:46:46.801655 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerName="registry-server" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.801663 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerName="registry-server" Nov 27 17:46:46 crc kubenswrapper[4809]: E1127 17:46:46.801675 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f21abe6-3443-4a72-b93b-574203c3f3b4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.801683 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f21abe6-3443-4a72-b93b-574203c3f3b4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 27 17:46:46 crc kubenswrapper[4809]: E1127 17:46:46.801701 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerName="extract-utilities" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.801707 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerName="extract-utilities" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.801942 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f21abe6-3443-4a72-b93b-574203c3f3b4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.801962 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a93d1ed-6bf0-46af-a351-fa067e36ad34" containerName="registry-server" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.802650 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.811027 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.811154 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.811199 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.812271 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.822899 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj"] Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.902684 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dngt\" (UniqueName: \"kubernetes.io/projected/e5b77adb-7f43-4df6-b676-cb82003887fc-kube-api-access-6dngt\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.903498 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:46 crc kubenswrapper[4809]: I1127 17:46:46.903798 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:47 crc kubenswrapper[4809]: I1127 17:46:47.005807 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:47 crc kubenswrapper[4809]: I1127 17:46:47.005880 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dngt\" (UniqueName: \"kubernetes.io/projected/e5b77adb-7f43-4df6-b676-cb82003887fc-kube-api-access-6dngt\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:47 crc kubenswrapper[4809]: I1127 17:46:47.006068 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:47 crc kubenswrapper[4809]: I1127 17:46:47.015656 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:47 crc kubenswrapper[4809]: I1127 17:46:47.015687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:47 crc kubenswrapper[4809]: I1127 17:46:47.024953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dngt\" (UniqueName: \"kubernetes.io/projected/e5b77adb-7f43-4df6-b676-cb82003887fc-kube-api-access-6dngt\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:47 crc kubenswrapper[4809]: I1127 17:46:47.119467 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:46:47 crc kubenswrapper[4809]: I1127 17:46:47.635202 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj"] Nov 27 17:46:47 crc kubenswrapper[4809]: I1127 17:46:47.751918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" event={"ID":"e5b77adb-7f43-4df6-b676-cb82003887fc","Type":"ContainerStarted","Data":"b70918ac8ae6b934ad12f445bfb50deb76fe44d5940375c6342290c4166e6084"} Nov 27 17:46:48 crc kubenswrapper[4809]: I1127 17:46:48.766641 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" event={"ID":"e5b77adb-7f43-4df6-b676-cb82003887fc","Type":"ContainerStarted","Data":"b281468552d1c31abb6d5154d3aea8f50da5f5630501523d9bac8dce093716ef"} Nov 27 17:46:48 crc kubenswrapper[4809]: I1127 17:46:48.793823 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" podStartSLOduration=2.275176241 podStartE2EDuration="2.793798947s" podCreationTimestamp="2025-11-27 17:46:46 +0000 UTC" firstStartedPulling="2025-11-27 17:46:47.638344248 +0000 UTC m=+2242.910801600" lastFinishedPulling="2025-11-27 17:46:48.156966954 +0000 UTC m=+2243.429424306" observedRunningTime="2025-11-27 17:46:48.784033991 +0000 UTC m=+2244.056491343" watchObservedRunningTime="2025-11-27 17:46:48.793798947 +0000 UTC m=+2244.066256299" Nov 27 17:47:25 crc kubenswrapper[4809]: I1127 17:47:25.804955 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xq9qw"] Nov 27 17:47:25 crc kubenswrapper[4809]: I1127 17:47:25.807684 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:25 crc kubenswrapper[4809]: I1127 17:47:25.817309 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xq9qw"] Nov 27 17:47:25 crc kubenswrapper[4809]: I1127 17:47:25.928946 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-catalog-content\") pod \"redhat-operators-xq9qw\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:25 crc kubenswrapper[4809]: I1127 17:47:25.929039 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc8k9\" (UniqueName: \"kubernetes.io/projected/10fee781-0481-49ab-a857-de91d23460a7-kube-api-access-qc8k9\") pod \"redhat-operators-xq9qw\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:25 crc kubenswrapper[4809]: I1127 17:47:25.929289 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-utilities\") pod \"redhat-operators-xq9qw\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:26 crc kubenswrapper[4809]: I1127 17:47:26.030705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-utilities\") pod \"redhat-operators-xq9qw\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:26 crc kubenswrapper[4809]: I1127 17:47:26.030913 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-catalog-content\") pod \"redhat-operators-xq9qw\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:26 crc kubenswrapper[4809]: I1127 17:47:26.030964 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc8k9\" (UniqueName: \"kubernetes.io/projected/10fee781-0481-49ab-a857-de91d23460a7-kube-api-access-qc8k9\") pod \"redhat-operators-xq9qw\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:26 crc kubenswrapper[4809]: I1127 17:47:26.031348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-utilities\") pod \"redhat-operators-xq9qw\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:26 crc kubenswrapper[4809]: I1127 17:47:26.031388 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-catalog-content\") pod \"redhat-operators-xq9qw\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:26 crc kubenswrapper[4809]: I1127 17:47:26.050902 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc8k9\" (UniqueName: \"kubernetes.io/projected/10fee781-0481-49ab-a857-de91d23460a7-kube-api-access-qc8k9\") pod \"redhat-operators-xq9qw\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:26 crc kubenswrapper[4809]: I1127 17:47:26.131703 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:26 crc kubenswrapper[4809]: I1127 17:47:26.630621 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xq9qw"] Nov 27 17:47:27 crc kubenswrapper[4809]: I1127 17:47:27.114221 4809 generic.go:334] "Generic (PLEG): container finished" podID="10fee781-0481-49ab-a857-de91d23460a7" containerID="a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135" exitCode=0 Nov 27 17:47:27 crc kubenswrapper[4809]: I1127 17:47:27.114271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq9qw" event={"ID":"10fee781-0481-49ab-a857-de91d23460a7","Type":"ContainerDied","Data":"a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135"} Nov 27 17:47:27 crc kubenswrapper[4809]: I1127 17:47:27.114300 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq9qw" event={"ID":"10fee781-0481-49ab-a857-de91d23460a7","Type":"ContainerStarted","Data":"ce7ff29e30c751700d6e85c0996a35a04edb8d84fdd07032b5e019a15fa15ca1"} Nov 27 17:47:28 crc kubenswrapper[4809]: I1127 17:47:28.124603 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq9qw" event={"ID":"10fee781-0481-49ab-a857-de91d23460a7","Type":"ContainerStarted","Data":"4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41"} Nov 27 17:47:29 crc kubenswrapper[4809]: I1127 17:47:29.136446 4809 generic.go:334] "Generic (PLEG): container finished" podID="10fee781-0481-49ab-a857-de91d23460a7" containerID="4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41" exitCode=0 Nov 27 17:47:29 crc kubenswrapper[4809]: I1127 17:47:29.136500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq9qw" event={"ID":"10fee781-0481-49ab-a857-de91d23460a7","Type":"ContainerDied","Data":"4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41"} Nov 27 17:47:31 crc kubenswrapper[4809]: I1127 17:47:31.159950 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq9qw" event={"ID":"10fee781-0481-49ab-a857-de91d23460a7","Type":"ContainerStarted","Data":"597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5"} Nov 27 17:47:31 crc kubenswrapper[4809]: I1127 17:47:31.187955 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xq9qw" podStartSLOduration=3.150136046 podStartE2EDuration="6.187922404s" podCreationTimestamp="2025-11-27 17:47:25 +0000 UTC" firstStartedPulling="2025-11-27 17:47:27.11644773 +0000 UTC m=+2282.388905082" lastFinishedPulling="2025-11-27 17:47:30.154234088 +0000 UTC m=+2285.426691440" observedRunningTime="2025-11-27 17:47:31.17967308 +0000 UTC m=+2286.452130442" watchObservedRunningTime="2025-11-27 17:47:31.187922404 +0000 UTC m=+2286.460379746" Nov 27 17:47:32 crc kubenswrapper[4809]: I1127 17:47:32.962237 4809 scope.go:117] "RemoveContainer" containerID="dc310c7bd30a10934a903eb14185ec56d7ab859fe90e16c215a545f292c5cc9c" Nov 27 17:47:32 crc kubenswrapper[4809]: I1127 17:47:32.984932 4809 scope.go:117] "RemoveContainer" containerID="2f12f08da0977ec4a82f97d6e7543e41c874274338cee8d4a212261510144e71" Nov 27 17:47:33 crc kubenswrapper[4809]: I1127 17:47:33.006386 4809 scope.go:117] "RemoveContainer" containerID="e0d64cac60b1d7dd70920d451042a26871b366b24598244ca13f72d1dea7dc9b" Nov 27 17:47:36 crc kubenswrapper[4809]: I1127 17:47:36.132726 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:36 crc kubenswrapper[4809]: I1127 17:47:36.133440 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:36 crc kubenswrapper[4809]: I1127 17:47:36.177120 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:36 crc kubenswrapper[4809]: I1127 17:47:36.249819 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:36 crc kubenswrapper[4809]: I1127 17:47:36.417507 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xq9qw"] Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.222310 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xq9qw" podUID="10fee781-0481-49ab-a857-de91d23460a7" containerName="registry-server" containerID="cri-o://597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5" gracePeriod=2 Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.695992 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.779121 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-catalog-content\") pod \"10fee781-0481-49ab-a857-de91d23460a7\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.779879 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc8k9\" (UniqueName: \"kubernetes.io/projected/10fee781-0481-49ab-a857-de91d23460a7-kube-api-access-qc8k9\") pod \"10fee781-0481-49ab-a857-de91d23460a7\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.779989 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-utilities\") pod \"10fee781-0481-49ab-a857-de91d23460a7\" (UID: \"10fee781-0481-49ab-a857-de91d23460a7\") " Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.781403 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-utilities" (OuterVolumeSpecName: "utilities") pod "10fee781-0481-49ab-a857-de91d23460a7" (UID: "10fee781-0481-49ab-a857-de91d23460a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.785616 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10fee781-0481-49ab-a857-de91d23460a7-kube-api-access-qc8k9" (OuterVolumeSpecName: "kube-api-access-qc8k9") pod "10fee781-0481-49ab-a857-de91d23460a7" (UID: "10fee781-0481-49ab-a857-de91d23460a7"). InnerVolumeSpecName "kube-api-access-qc8k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.881677 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.881710 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc8k9\" (UniqueName: \"kubernetes.io/projected/10fee781-0481-49ab-a857-de91d23460a7-kube-api-access-qc8k9\") on node \"crc\" DevicePath \"\"" Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.887830 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10fee781-0481-49ab-a857-de91d23460a7" (UID: "10fee781-0481-49ab-a857-de91d23460a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:47:38 crc kubenswrapper[4809]: I1127 17:47:38.984087 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10fee781-0481-49ab-a857-de91d23460a7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.235014 4809 generic.go:334] "Generic (PLEG): container finished" podID="10fee781-0481-49ab-a857-de91d23460a7" containerID="597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5" exitCode=0 Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.235057 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq9qw" event={"ID":"10fee781-0481-49ab-a857-de91d23460a7","Type":"ContainerDied","Data":"597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5"} Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.235093 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq9qw" event={"ID":"10fee781-0481-49ab-a857-de91d23460a7","Type":"ContainerDied","Data":"ce7ff29e30c751700d6e85c0996a35a04edb8d84fdd07032b5e019a15fa15ca1"} Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.235107 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xq9qw" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.235135 4809 scope.go:117] "RemoveContainer" containerID="597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.259347 4809 scope.go:117] "RemoveContainer" containerID="4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.269547 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xq9qw"] Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.280238 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xq9qw"] Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.294651 4809 scope.go:117] "RemoveContainer" containerID="a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.334905 4809 scope.go:117] "RemoveContainer" containerID="597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5" Nov 27 17:47:39 crc kubenswrapper[4809]: E1127 17:47:39.335420 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5\": container with ID starting with 597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5 not found: ID does not exist" containerID="597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.335463 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5"} err="failed to get container status \"597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5\": rpc error: code = NotFound desc = could not find container \"597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5\": container with ID starting with 597285ee8a30b0524b56ffbaf8ff08bf4b01522a807b9d3034990b38d53b2bc5 not found: ID does not exist" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.335495 4809 scope.go:117] "RemoveContainer" containerID="4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41" Nov 27 17:47:39 crc kubenswrapper[4809]: E1127 17:47:39.335972 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41\": container with ID starting with 4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41 not found: ID does not exist" containerID="4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.336020 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41"} err="failed to get container status \"4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41\": rpc error: code = NotFound desc = could not find container \"4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41\": container with ID starting with 4716226e3a7c18504510d857a17e48fcac8a63598b072ed02a9da680df02de41 not found: ID does not exist" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.336049 4809 scope.go:117] "RemoveContainer" containerID="a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135" Nov 27 17:47:39 crc kubenswrapper[4809]: E1127 17:47:39.336472 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135\": container with ID starting with a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135 not found: ID does not exist" containerID="a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.336500 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135"} err="failed to get container status \"a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135\": rpc error: code = NotFound desc = could not find container \"a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135\": container with ID starting with a45858f137530aa17432f98a8e6f59c5a4a006747dc82f969b94ba524e385135 not found: ID does not exist" Nov 27 17:47:39 crc kubenswrapper[4809]: I1127 17:47:39.470010 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10fee781-0481-49ab-a857-de91d23460a7" path="/var/lib/kubelet/pods/10fee781-0481-49ab-a857-de91d23460a7/volumes" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.604202 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zjpbq"] Nov 27 17:47:51 crc kubenswrapper[4809]: E1127 17:47:51.605341 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10fee781-0481-49ab-a857-de91d23460a7" containerName="extract-content" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.605359 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="10fee781-0481-49ab-a857-de91d23460a7" containerName="extract-content" Nov 27 17:47:51 crc kubenswrapper[4809]: E1127 17:47:51.605399 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10fee781-0481-49ab-a857-de91d23460a7" containerName="extract-utilities" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.605407 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="10fee781-0481-49ab-a857-de91d23460a7" containerName="extract-utilities" Nov 27 17:47:51 crc kubenswrapper[4809]: E1127 17:47:51.605430 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10fee781-0481-49ab-a857-de91d23460a7" containerName="registry-server" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.605438 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="10fee781-0481-49ab-a857-de91d23460a7" containerName="registry-server" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.605668 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="10fee781-0481-49ab-a857-de91d23460a7" containerName="registry-server" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.607489 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.617361 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjpbq"] Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.656086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-catalog-content\") pod \"redhat-marketplace-zjpbq\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.656208 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-utilities\") pod \"redhat-marketplace-zjpbq\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.656247 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxcms\" (UniqueName: \"kubernetes.io/projected/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-kube-api-access-lxcms\") pod \"redhat-marketplace-zjpbq\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.758251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-utilities\") pod \"redhat-marketplace-zjpbq\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.758308 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxcms\" (UniqueName: \"kubernetes.io/projected/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-kube-api-access-lxcms\") pod \"redhat-marketplace-zjpbq\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.758490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-catalog-content\") pod \"redhat-marketplace-zjpbq\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.759011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-catalog-content\") pod \"redhat-marketplace-zjpbq\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.759265 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-utilities\") pod \"redhat-marketplace-zjpbq\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.781909 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxcms\" (UniqueName: \"kubernetes.io/projected/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-kube-api-access-lxcms\") pod \"redhat-marketplace-zjpbq\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:51 crc kubenswrapper[4809]: I1127 17:47:51.930205 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:47:52 crc kubenswrapper[4809]: I1127 17:47:52.385282 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjpbq"] Nov 27 17:47:53 crc kubenswrapper[4809]: I1127 17:47:53.373603 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerID="b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442" exitCode=0 Nov 27 17:47:53 crc kubenswrapper[4809]: I1127 17:47:53.373713 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjpbq" event={"ID":"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa","Type":"ContainerDied","Data":"b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442"} Nov 27 17:47:53 crc kubenswrapper[4809]: I1127 17:47:53.374239 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjpbq" event={"ID":"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa","Type":"ContainerStarted","Data":"a3fee7d22c27590d3e58a023ad61fb1590375342fe22c3c86ff5efdd05a3949a"} Nov 27 17:47:54 crc kubenswrapper[4809]: I1127 17:47:54.388562 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerID="2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659" exitCode=0 Nov 27 17:47:54 crc kubenswrapper[4809]: I1127 17:47:54.388658 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjpbq" event={"ID":"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa","Type":"ContainerDied","Data":"2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659"} Nov 27 17:47:55 crc kubenswrapper[4809]: I1127 17:47:55.401380 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjpbq" event={"ID":"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa","Type":"ContainerStarted","Data":"ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e"} Nov 27 17:48:01 crc kubenswrapper[4809]: I1127 17:48:01.930428 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:48:01 crc kubenswrapper[4809]: I1127 17:48:01.931052 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:48:01 crc kubenswrapper[4809]: I1127 17:48:01.982362 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:48:02 crc kubenswrapper[4809]: I1127 17:48:02.007822 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zjpbq" podStartSLOduration=9.27193656 podStartE2EDuration="11.007798324s" podCreationTimestamp="2025-11-27 17:47:51 +0000 UTC" firstStartedPulling="2025-11-27 17:47:53.37709496 +0000 UTC m=+2308.649552312" lastFinishedPulling="2025-11-27 17:47:55.112956724 +0000 UTC m=+2310.385414076" observedRunningTime="2025-11-27 17:47:55.424421117 +0000 UTC m=+2310.696878479" watchObservedRunningTime="2025-11-27 17:48:02.007798324 +0000 UTC m=+2317.280255686" Nov 27 17:48:02 crc kubenswrapper[4809]: I1127 17:48:02.520027 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:48:02 crc kubenswrapper[4809]: I1127 17:48:02.574299 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjpbq"] Nov 27 17:48:04 crc kubenswrapper[4809]: I1127 17:48:04.492507 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zjpbq" podUID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerName="registry-server" containerID="cri-o://ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e" gracePeriod=2 Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.067406 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.147670 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxcms\" (UniqueName: \"kubernetes.io/projected/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-kube-api-access-lxcms\") pod \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.147822 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-utilities\") pod \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.147962 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-catalog-content\") pod \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\" (UID: \"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa\") " Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.148825 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-utilities" (OuterVolumeSpecName: "utilities") pod "fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" (UID: "fd5fdcc3-5498-4d11-b889-06ab10c6c6fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.154178 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-kube-api-access-lxcms" (OuterVolumeSpecName: "kube-api-access-lxcms") pod "fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" (UID: "fd5fdcc3-5498-4d11-b889-06ab10c6c6fa"). InnerVolumeSpecName "kube-api-access-lxcms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.167232 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" (UID: "fd5fdcc3-5498-4d11-b889-06ab10c6c6fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.252004 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxcms\" (UniqueName: \"kubernetes.io/projected/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-kube-api-access-lxcms\") on node \"crc\" DevicePath \"\"" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.252598 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.252635 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.507712 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerID="ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e" exitCode=0 Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.507777 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjpbq" event={"ID":"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa","Type":"ContainerDied","Data":"ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e"} Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.507838 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjpbq" event={"ID":"fd5fdcc3-5498-4d11-b889-06ab10c6c6fa","Type":"ContainerDied","Data":"a3fee7d22c27590d3e58a023ad61fb1590375342fe22c3c86ff5efdd05a3949a"} Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.507857 4809 scope.go:117] "RemoveContainer" containerID="ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.507795 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjpbq" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.531923 4809 scope.go:117] "RemoveContainer" containerID="2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.541292 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjpbq"] Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.553141 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjpbq"] Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.571387 4809 scope.go:117] "RemoveContainer" containerID="b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.614274 4809 scope.go:117] "RemoveContainer" containerID="ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e" Nov 27 17:48:05 crc kubenswrapper[4809]: E1127 17:48:05.614725 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e\": container with ID starting with ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e not found: ID does not exist" containerID="ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.614798 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e"} err="failed to get container status \"ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e\": rpc error: code = NotFound desc = could not find container \"ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e\": container with ID starting with ac3a5e47c20863b9e4b199002945fb27656ad1338f63afdc4daaa0d34822981e not found: ID does not exist" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.614836 4809 scope.go:117] "RemoveContainer" containerID="2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659" Nov 27 17:48:05 crc kubenswrapper[4809]: E1127 17:48:05.615298 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659\": container with ID starting with 2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659 not found: ID does not exist" containerID="2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.615348 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659"} err="failed to get container status \"2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659\": rpc error: code = NotFound desc = could not find container \"2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659\": container with ID starting with 2d0bdc54f996c0bd020429f2a29bc3a0ef5d2273d4737658b0248886fd30f659 not found: ID does not exist" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.615381 4809 scope.go:117] "RemoveContainer" containerID="b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442" Nov 27 17:48:05 crc kubenswrapper[4809]: E1127 17:48:05.615692 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442\": container with ID starting with b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442 not found: ID does not exist" containerID="b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442" Nov 27 17:48:05 crc kubenswrapper[4809]: I1127 17:48:05.615728 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442"} err="failed to get container status \"b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442\": rpc error: code = NotFound desc = could not find container \"b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442\": container with ID starting with b88b4f8b57ca959028b6248b65f1312a27f7f0f0b88daecbe7d7279f332b3442 not found: ID does not exist" Nov 27 17:48:07 crc kubenswrapper[4809]: I1127 17:48:07.470096 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" path="/var/lib/kubelet/pods/fd5fdcc3-5498-4d11-b889-06ab10c6c6fa/volumes" Nov 27 17:48:13 crc kubenswrapper[4809]: I1127 17:48:13.038277 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ql6vt"] Nov 27 17:48:13 crc kubenswrapper[4809]: I1127 17:48:13.047434 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-54b8-account-create-update-4t2zg"] Nov 27 17:48:13 crc kubenswrapper[4809]: I1127 17:48:13.057074 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ql6vt"] Nov 27 17:48:13 crc kubenswrapper[4809]: I1127 17:48:13.065884 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-54b8-account-create-update-4t2zg"] Nov 27 17:48:13 crc kubenswrapper[4809]: I1127 17:48:13.468918 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f" path="/var/lib/kubelet/pods/16f2aca6-31a1-41d6-9fbb-1a3fd8ff3f8f/volumes" Nov 27 17:48:13 crc kubenswrapper[4809]: I1127 17:48:13.469702 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f8d060-b713-4586-b388-c68c51647297" path="/var/lib/kubelet/pods/91f8d060-b713-4586-b388-c68c51647297/volumes" Nov 27 17:48:17 crc kubenswrapper[4809]: I1127 17:48:17.028301 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-ldz52"] Nov 27 17:48:17 crc kubenswrapper[4809]: I1127 17:48:17.039974 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e5a0-account-create-update-px7dm"] Nov 27 17:48:17 crc kubenswrapper[4809]: I1127 17:48:17.048520 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e5a0-account-create-update-px7dm"] Nov 27 17:48:17 crc kubenswrapper[4809]: I1127 17:48:17.057217 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-ldz52"] Nov 27 17:48:17 crc kubenswrapper[4809]: I1127 17:48:17.472229 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61" path="/var/lib/kubelet/pods/2a54a5ce-419a-4b39-9fc9-cd80dd4c1e61/volumes" Nov 27 17:48:17 crc kubenswrapper[4809]: I1127 17:48:17.472854 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9da48a67-65c8-462f-91af-ed6c4192010d" path="/var/lib/kubelet/pods/9da48a67-65c8-462f-91af-ed6c4192010d/volumes" Nov 27 17:48:18 crc kubenswrapper[4809]: I1127 17:48:18.028205 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-8e30-account-create-update-b8pvk"] Nov 27 17:48:18 crc kubenswrapper[4809]: I1127 17:48:18.037439 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-vqr6f"] Nov 27 17:48:18 crc kubenswrapper[4809]: I1127 17:48:18.047832 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-vqr6f"] Nov 27 17:48:18 crc kubenswrapper[4809]: I1127 17:48:18.058345 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-8e30-account-create-update-b8pvk"] Nov 27 17:48:19 crc kubenswrapper[4809]: I1127 17:48:19.469685 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b9207d8-29be-4fa6-bd67-04eb1afc29a7" path="/var/lib/kubelet/pods/5b9207d8-29be-4fa6-bd67-04eb1afc29a7/volumes" Nov 27 17:48:19 crc kubenswrapper[4809]: I1127 17:48:19.470676 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61639b29-dc56-47d2-8c8d-44aea5ba45f7" path="/var/lib/kubelet/pods/61639b29-dc56-47d2-8c8d-44aea5ba45f7/volumes" Nov 27 17:48:22 crc kubenswrapper[4809]: I1127 17:48:22.661600 4809 generic.go:334] "Generic (PLEG): container finished" podID="e5b77adb-7f43-4df6-b676-cb82003887fc" containerID="b281468552d1c31abb6d5154d3aea8f50da5f5630501523d9bac8dce093716ef" exitCode=0 Nov 27 17:48:22 crc kubenswrapper[4809]: I1127 17:48:22.661680 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" event={"ID":"e5b77adb-7f43-4df6-b676-cb82003887fc","Type":"ContainerDied","Data":"b281468552d1c31abb6d5154d3aea8f50da5f5630501523d9bac8dce093716ef"} Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.093252 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.236548 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-ssh-key\") pod \"e5b77adb-7f43-4df6-b676-cb82003887fc\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.236688 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dngt\" (UniqueName: \"kubernetes.io/projected/e5b77adb-7f43-4df6-b676-cb82003887fc-kube-api-access-6dngt\") pod \"e5b77adb-7f43-4df6-b676-cb82003887fc\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.237437 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-inventory\") pod \"e5b77adb-7f43-4df6-b676-cb82003887fc\" (UID: \"e5b77adb-7f43-4df6-b676-cb82003887fc\") " Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.244037 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5b77adb-7f43-4df6-b676-cb82003887fc-kube-api-access-6dngt" (OuterVolumeSpecName: "kube-api-access-6dngt") pod "e5b77adb-7f43-4df6-b676-cb82003887fc" (UID: "e5b77adb-7f43-4df6-b676-cb82003887fc"). InnerVolumeSpecName "kube-api-access-6dngt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.266940 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e5b77adb-7f43-4df6-b676-cb82003887fc" (UID: "e5b77adb-7f43-4df6-b676-cb82003887fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.268500 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-inventory" (OuterVolumeSpecName: "inventory") pod "e5b77adb-7f43-4df6-b676-cb82003887fc" (UID: "e5b77adb-7f43-4df6-b676-cb82003887fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.340757 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.340803 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dngt\" (UniqueName: \"kubernetes.io/projected/e5b77adb-7f43-4df6-b676-cb82003887fc-kube-api-access-6dngt\") on node \"crc\" DevicePath \"\"" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.340818 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5b77adb-7f43-4df6-b676-cb82003887fc-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.684666 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" event={"ID":"e5b77adb-7f43-4df6-b676-cb82003887fc","Type":"ContainerDied","Data":"b70918ac8ae6b934ad12f445bfb50deb76fe44d5940375c6342290c4166e6084"} Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.684706 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b70918ac8ae6b934ad12f445bfb50deb76fe44d5940375c6342290c4166e6084" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.684782 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.776375 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4"] Nov 27 17:48:24 crc kubenswrapper[4809]: E1127 17:48:24.777865 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerName="extract-utilities" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.777890 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerName="extract-utilities" Nov 27 17:48:24 crc kubenswrapper[4809]: E1127 17:48:24.777904 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerName="extract-content" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.777911 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerName="extract-content" Nov 27 17:48:24 crc kubenswrapper[4809]: E1127 17:48:24.777937 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerName="registry-server" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.777944 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerName="registry-server" Nov 27 17:48:24 crc kubenswrapper[4809]: E1127 17:48:24.777981 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b77adb-7f43-4df6-b676-cb82003887fc" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.777989 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b77adb-7f43-4df6-b676-cb82003887fc" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.778199 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd5fdcc3-5498-4d11-b889-06ab10c6c6fa" containerName="registry-server" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.778229 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5b77adb-7f43-4df6-b676-cb82003887fc" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.779101 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.782278 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.782344 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.782484 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.782586 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.788194 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4"] Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.952165 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-78ns4\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.952453 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqmjh\" (UniqueName: \"kubernetes.io/projected/2021472c-275a-44fd-802b-37e72374a72d-kube-api-access-kqmjh\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-78ns4\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:24 crc kubenswrapper[4809]: I1127 17:48:24.952543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-78ns4\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.054329 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqmjh\" (UniqueName: \"kubernetes.io/projected/2021472c-275a-44fd-802b-37e72374a72d-kube-api-access-kqmjh\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-78ns4\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.054395 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-78ns4\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.054575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-78ns4\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.058696 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-78ns4\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.059296 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-78ns4\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.085813 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqmjh\" (UniqueName: \"kubernetes.io/projected/2021472c-275a-44fd-802b-37e72374a72d-kube-api-access-kqmjh\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-78ns4\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.097025 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.779940 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.780496 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.815550 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4"] Nov 27 17:48:25 crc kubenswrapper[4809]: W1127 17:48:25.818851 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2021472c_275a_44fd_802b_37e72374a72d.slice/crio-b0da2fe936ff7bb4acd2ff9447f5fe959d81edf6008314c0a3cc92d019ba4c98 WatchSource:0}: Error finding container b0da2fe936ff7bb4acd2ff9447f5fe959d81edf6008314c0a3cc92d019ba4c98: Status 404 returned error can't find the container with id b0da2fe936ff7bb4acd2ff9447f5fe959d81edf6008314c0a3cc92d019ba4c98 Nov 27 17:48:25 crc kubenswrapper[4809]: I1127 17:48:25.822750 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 17:48:26 crc kubenswrapper[4809]: I1127 17:48:26.252192 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:48:26 crc kubenswrapper[4809]: I1127 17:48:26.703060 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" event={"ID":"2021472c-275a-44fd-802b-37e72374a72d","Type":"ContainerStarted","Data":"82a8c46ddd43d13ea1c051fc3b5ab78a45bf61bb2884c1b4758b6c6ee397add2"} Nov 27 17:48:26 crc kubenswrapper[4809]: I1127 17:48:26.703421 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" event={"ID":"2021472c-275a-44fd-802b-37e72374a72d","Type":"ContainerStarted","Data":"b0da2fe936ff7bb4acd2ff9447f5fe959d81edf6008314c0a3cc92d019ba4c98"} Nov 27 17:48:26 crc kubenswrapper[4809]: I1127 17:48:26.722036 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" podStartSLOduration=2.2942897159999998 podStartE2EDuration="2.722016269s" podCreationTimestamp="2025-11-27 17:48:24 +0000 UTC" firstStartedPulling="2025-11-27 17:48:25.822478952 +0000 UTC m=+2341.094936304" lastFinishedPulling="2025-11-27 17:48:26.250205505 +0000 UTC m=+2341.522662857" observedRunningTime="2025-11-27 17:48:26.716100198 +0000 UTC m=+2341.988557550" watchObservedRunningTime="2025-11-27 17:48:26.722016269 +0000 UTC m=+2341.994473621" Nov 27 17:48:33 crc kubenswrapper[4809]: I1127 17:48:33.121097 4809 scope.go:117] "RemoveContainer" containerID="aba6cb2b85c552c5ad546aeb8dba03d515379e238c98345f5309dfd8081cca65" Nov 27 17:48:33 crc kubenswrapper[4809]: I1127 17:48:33.145039 4809 scope.go:117] "RemoveContainer" containerID="f61380d9bd12c2bcdd88414cdbf980b53d492a95de6531d04ec443c17f571c72" Nov 27 17:48:33 crc kubenswrapper[4809]: I1127 17:48:33.201365 4809 scope.go:117] "RemoveContainer" containerID="478b94db3d1c86917b78ae47a4fa5c9dbfcdd36b291df6e8312989a838f2370c" Nov 27 17:48:33 crc kubenswrapper[4809]: I1127 17:48:33.243938 4809 scope.go:117] "RemoveContainer" containerID="76b55033f40a72810e20207f7f86639cf89daa9f50675468672928fb92f8d130" Nov 27 17:48:33 crc kubenswrapper[4809]: I1127 17:48:33.297023 4809 scope.go:117] "RemoveContainer" containerID="8c59fd8c4df8f089ddbfa38ce89e026badace6a1d40a425fd77c62ab417cab93" Nov 27 17:48:33 crc kubenswrapper[4809]: I1127 17:48:33.357044 4809 scope.go:117] "RemoveContainer" containerID="300d34c50c0b5b59ec864c21ee6e262af11ca7a6702fb6464dccf815090c3bf4" Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.034498 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-87f3-account-create-update-klgkz"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.047071 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-8bnvg"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.058394 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-7fd4-account-create-update-pg2kx"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.067363 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-vtslx"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.076702 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-87f3-account-create-update-klgkz"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.085276 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vg6fv"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.093461 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-8bnvg"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.101895 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-7fd4-account-create-update-pg2kx"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.110779 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-vtslx"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.118999 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vg6fv"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.126788 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-19b4-account-create-update-n9t2g"] Nov 27 17:48:36 crc kubenswrapper[4809]: I1127 17:48:36.134616 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-19b4-account-create-update-n9t2g"] Nov 27 17:48:37 crc kubenswrapper[4809]: I1127 17:48:37.468080 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="025088fd-c5dc-452e-aceb-c4a04cc09033" path="/var/lib/kubelet/pods/025088fd-c5dc-452e-aceb-c4a04cc09033/volumes" Nov 27 17:48:37 crc kubenswrapper[4809]: I1127 17:48:37.468966 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ae75b87-925b-4b0c-9a59-1d59f534f9f2" path="/var/lib/kubelet/pods/4ae75b87-925b-4b0c-9a59-1d59f534f9f2/volumes" Nov 27 17:48:37 crc kubenswrapper[4809]: I1127 17:48:37.469577 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dcd2971-5635-42ab-af5c-ee867a78bfa1" path="/var/lib/kubelet/pods/4dcd2971-5635-42ab-af5c-ee867a78bfa1/volumes" Nov 27 17:48:37 crc kubenswrapper[4809]: I1127 17:48:37.470173 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b374716e-7e54-4be1-96bf-3c9e9bd379b8" path="/var/lib/kubelet/pods/b374716e-7e54-4be1-96bf-3c9e9bd379b8/volumes" Nov 27 17:48:37 crc kubenswrapper[4809]: I1127 17:48:37.471267 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4a9f9d1-a45a-49a1-95bd-1681a1b220ff" path="/var/lib/kubelet/pods/c4a9f9d1-a45a-49a1-95bd-1681a1b220ff/volumes" Nov 27 17:48:37 crc kubenswrapper[4809]: I1127 17:48:37.471828 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d44f63e5-7e83-4cd8-a488-0e88f7635938" path="/var/lib/kubelet/pods/d44f63e5-7e83-4cd8-a488-0e88f7635938/volumes" Nov 27 17:48:43 crc kubenswrapper[4809]: I1127 17:48:43.039496 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-bm4mw"] Nov 27 17:48:43 crc kubenswrapper[4809]: I1127 17:48:43.048759 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-bm4mw"] Nov 27 17:48:43 crc kubenswrapper[4809]: I1127 17:48:43.469706 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acc5484a-4058-4a5d-97a0-e84aaa75905f" path="/var/lib/kubelet/pods/acc5484a-4058-4a5d-97a0-e84aaa75905f/volumes" Nov 27 17:48:45 crc kubenswrapper[4809]: I1127 17:48:45.026711 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-5tlhg"] Nov 27 17:48:45 crc kubenswrapper[4809]: I1127 17:48:45.034880 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-5tlhg"] Nov 27 17:48:45 crc kubenswrapper[4809]: I1127 17:48:45.470034 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4891ffad-a908-49ac-97f5-196ae7a2e919" path="/var/lib/kubelet/pods/4891ffad-a908-49ac-97f5-196ae7a2e919/volumes" Nov 27 17:48:55 crc kubenswrapper[4809]: I1127 17:48:55.780076 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:48:55 crc kubenswrapper[4809]: I1127 17:48:55.780558 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:49:19 crc kubenswrapper[4809]: I1127 17:49:19.041942 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-k747n"] Nov 27 17:49:19 crc kubenswrapper[4809]: I1127 17:49:19.057544 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ktqz6"] Nov 27 17:49:19 crc kubenswrapper[4809]: I1127 17:49:19.069398 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-rhmfz"] Nov 27 17:49:19 crc kubenswrapper[4809]: I1127 17:49:19.079690 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-rhmfz"] Nov 27 17:49:19 crc kubenswrapper[4809]: I1127 17:49:19.088791 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-k747n"] Nov 27 17:49:19 crc kubenswrapper[4809]: I1127 17:49:19.097888 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ktqz6"] Nov 27 17:49:19 crc kubenswrapper[4809]: I1127 17:49:19.470290 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bbb5fa1-67a7-4d6e-9354-36598e97d9d0" path="/var/lib/kubelet/pods/0bbb5fa1-67a7-4d6e-9354-36598e97d9d0/volumes" Nov 27 17:49:19 crc kubenswrapper[4809]: I1127 17:49:19.470932 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e177c61-cc3e-444a-b6ed-aecd6d262b7c" path="/var/lib/kubelet/pods/4e177c61-cc3e-444a-b6ed-aecd6d262b7c/volumes" Nov 27 17:49:19 crc kubenswrapper[4809]: I1127 17:49:19.471609 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b62f0e-7576-4e57-9993-30000f0eee5a" path="/var/lib/kubelet/pods/57b62f0e-7576-4e57-9993-30000f0eee5a/volumes" Nov 27 17:49:20 crc kubenswrapper[4809]: I1127 17:49:20.030272 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-97thg"] Nov 27 17:49:20 crc kubenswrapper[4809]: I1127 17:49:20.040853 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-97thg"] Nov 27 17:49:21 crc kubenswrapper[4809]: I1127 17:49:21.469052 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c184a0c3-5bb0-49f0-966f-fb565fd13202" path="/var/lib/kubelet/pods/c184a0c3-5bb0-49f0-966f-fb565fd13202/volumes" Nov 27 17:49:25 crc kubenswrapper[4809]: I1127 17:49:25.779617 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:49:25 crc kubenswrapper[4809]: I1127 17:49:25.780227 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:49:25 crc kubenswrapper[4809]: I1127 17:49:25.780274 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:49:25 crc kubenswrapper[4809]: I1127 17:49:25.781054 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:49:25 crc kubenswrapper[4809]: I1127 17:49:25.781109 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" gracePeriod=600 Nov 27 17:49:25 crc kubenswrapper[4809]: E1127 17:49:25.902163 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:49:26 crc kubenswrapper[4809]: I1127 17:49:26.262391 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" exitCode=0 Nov 27 17:49:26 crc kubenswrapper[4809]: I1127 17:49:26.262717 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0"} Nov 27 17:49:26 crc kubenswrapper[4809]: I1127 17:49:26.262769 4809 scope.go:117] "RemoveContainer" containerID="9a6ef73344f0f4b1da4284fd6ac88d1404b7c5e1410dce4a9042ec288f5d1365" Nov 27 17:49:26 crc kubenswrapper[4809]: I1127 17:49:26.263390 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:49:26 crc kubenswrapper[4809]: E1127 17:49:26.263634 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.554455 4809 scope.go:117] "RemoveContainer" containerID="add867ab2e890587644a3fa7893d6444487f2b07f654196452bfdcc21c669ea8" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.586601 4809 scope.go:117] "RemoveContainer" containerID="800ddd6ba7a20a24f4be54afcd2a5622e7e8ccf7e7b8538fc7fc014b6e666719" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.631817 4809 scope.go:117] "RemoveContainer" containerID="6de46e7d9a5928fd2d465277012c3423b9e297c739d698796bfebc205c6b9091" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.696610 4809 scope.go:117] "RemoveContainer" containerID="09d8633d799bef52f0183e342dc16ab5245ee464bc71186d4d6914dddc30cbc2" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.724506 4809 scope.go:117] "RemoveContainer" containerID="aa81b99ab71174dca079f8cce2d31038d0c85915e7e7aa876ddb7b5d15cddbab" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.778296 4809 scope.go:117] "RemoveContainer" containerID="b52055a16d85e6af2503992f7364d1eb40fb9f67870308ae59d2dc63673b2fd0" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.830724 4809 scope.go:117] "RemoveContainer" containerID="6b30ff49af31a2664b0b05e11743403530b80d1a8f2c1bb00688074ae847659e" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.860698 4809 scope.go:117] "RemoveContainer" containerID="7f49e17057866859ef63358c523acf92a4e1b191d6b839043accc88ec8db198c" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.885837 4809 scope.go:117] "RemoveContainer" containerID="2c83e01330814afae94a2e95302d070a959f576dc9b7f275f32f672d4d8dd57f" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.909232 4809 scope.go:117] "RemoveContainer" containerID="f65f913622348400865367354a71250bcd9fe05e5fe48157f3efa94651a00405" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.936315 4809 scope.go:117] "RemoveContainer" containerID="186036199ff8fe73257a5d42a2cbabb177d45c47333b1d8d3c20053c2fd4e047" Nov 27 17:49:33 crc kubenswrapper[4809]: I1127 17:49:33.961047 4809 scope.go:117] "RemoveContainer" containerID="a2ef4cd023c1bc6fc21cd49a10d176140642e6687e16293b03942033304efb52" Nov 27 17:49:38 crc kubenswrapper[4809]: I1127 17:49:38.458002 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:49:38 crc kubenswrapper[4809]: E1127 17:49:38.458875 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:49:40 crc kubenswrapper[4809]: I1127 17:49:40.397502 4809 generic.go:334] "Generic (PLEG): container finished" podID="2021472c-275a-44fd-802b-37e72374a72d" containerID="82a8c46ddd43d13ea1c051fc3b5ab78a45bf61bb2884c1b4758b6c6ee397add2" exitCode=0 Nov 27 17:49:40 crc kubenswrapper[4809]: I1127 17:49:40.397835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" event={"ID":"2021472c-275a-44fd-802b-37e72374a72d","Type":"ContainerDied","Data":"82a8c46ddd43d13ea1c051fc3b5ab78a45bf61bb2884c1b4758b6c6ee397add2"} Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.036673 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-t2gnx"] Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.045183 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-t2gnx"] Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.471852 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a3b3964-4402-49a4-86ba-e6e7909082ec" path="/var/lib/kubelet/pods/9a3b3964-4402-49a4-86ba-e6e7909082ec/volumes" Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.827734 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.957463 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-inventory\") pod \"2021472c-275a-44fd-802b-37e72374a72d\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.957673 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-ssh-key\") pod \"2021472c-275a-44fd-802b-37e72374a72d\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.957761 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqmjh\" (UniqueName: \"kubernetes.io/projected/2021472c-275a-44fd-802b-37e72374a72d-kube-api-access-kqmjh\") pod \"2021472c-275a-44fd-802b-37e72374a72d\" (UID: \"2021472c-275a-44fd-802b-37e72374a72d\") " Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.964140 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2021472c-275a-44fd-802b-37e72374a72d-kube-api-access-kqmjh" (OuterVolumeSpecName: "kube-api-access-kqmjh") pod "2021472c-275a-44fd-802b-37e72374a72d" (UID: "2021472c-275a-44fd-802b-37e72374a72d"). InnerVolumeSpecName "kube-api-access-kqmjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.988510 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2021472c-275a-44fd-802b-37e72374a72d" (UID: "2021472c-275a-44fd-802b-37e72374a72d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:49:41 crc kubenswrapper[4809]: I1127 17:49:41.991180 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-inventory" (OuterVolumeSpecName: "inventory") pod "2021472c-275a-44fd-802b-37e72374a72d" (UID: "2021472c-275a-44fd-802b-37e72374a72d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.060585 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.060636 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqmjh\" (UniqueName: \"kubernetes.io/projected/2021472c-275a-44fd-802b-37e72374a72d-kube-api-access-kqmjh\") on node \"crc\" DevicePath \"\"" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.060651 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2021472c-275a-44fd-802b-37e72374a72d-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.416447 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" event={"ID":"2021472c-275a-44fd-802b-37e72374a72d","Type":"ContainerDied","Data":"b0da2fe936ff7bb4acd2ff9447f5fe959d81edf6008314c0a3cc92d019ba4c98"} Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.416485 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0da2fe936ff7bb4acd2ff9447f5fe959d81edf6008314c0a3cc92d019ba4c98" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.416486 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-78ns4" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.522956 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7"] Nov 27 17:49:42 crc kubenswrapper[4809]: E1127 17:49:42.523518 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2021472c-275a-44fd-802b-37e72374a72d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.523535 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2021472c-275a-44fd-802b-37e72374a72d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.523930 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2021472c-275a-44fd-802b-37e72374a72d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.524610 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.526554 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.528609 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.528881 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.529417 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.538086 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7"] Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.671098 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nb2g\" (UniqueName: \"kubernetes.io/projected/c91fe577-8e90-411e-b078-8ece235f2f93-kube-api-access-6nb2g\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.671486 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.671546 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.773833 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.773947 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.774055 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nb2g\" (UniqueName: \"kubernetes.io/projected/c91fe577-8e90-411e-b078-8ece235f2f93-kube-api-access-6nb2g\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.780758 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.780851 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.795806 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nb2g\" (UniqueName: \"kubernetes.io/projected/c91fe577-8e90-411e-b078-8ece235f2f93-kube-api-access-6nb2g\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:42 crc kubenswrapper[4809]: I1127 17:49:42.843815 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:43 crc kubenswrapper[4809]: I1127 17:49:43.338391 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7"] Nov 27 17:49:43 crc kubenswrapper[4809]: I1127 17:49:43.426404 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" event={"ID":"c91fe577-8e90-411e-b078-8ece235f2f93","Type":"ContainerStarted","Data":"780dd44306bd10dab3425feacc951e67b671d17c64904bd8d286fd4247bc1050"} Nov 27 17:49:44 crc kubenswrapper[4809]: I1127 17:49:44.436843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" event={"ID":"c91fe577-8e90-411e-b078-8ece235f2f93","Type":"ContainerStarted","Data":"b67e23f70f4533511e09fac2edd5683fa3b7f8fdb93aec2eb45cb6e12b33f26b"} Nov 27 17:49:44 crc kubenswrapper[4809]: I1127 17:49:44.470874 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" podStartSLOduration=1.879037152 podStartE2EDuration="2.470858409s" podCreationTimestamp="2025-11-27 17:49:42 +0000 UTC" firstStartedPulling="2025-11-27 17:49:43.348041822 +0000 UTC m=+2418.620499184" lastFinishedPulling="2025-11-27 17:49:43.939863089 +0000 UTC m=+2419.212320441" observedRunningTime="2025-11-27 17:49:44.469367369 +0000 UTC m=+2419.741824721" watchObservedRunningTime="2025-11-27 17:49:44.470858409 +0000 UTC m=+2419.743315761" Nov 27 17:49:49 crc kubenswrapper[4809]: I1127 17:49:49.458358 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:49:49 crc kubenswrapper[4809]: E1127 17:49:49.459223 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:49:49 crc kubenswrapper[4809]: I1127 17:49:49.478715 4809 generic.go:334] "Generic (PLEG): container finished" podID="c91fe577-8e90-411e-b078-8ece235f2f93" containerID="b67e23f70f4533511e09fac2edd5683fa3b7f8fdb93aec2eb45cb6e12b33f26b" exitCode=0 Nov 27 17:49:49 crc kubenswrapper[4809]: I1127 17:49:49.478774 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" event={"ID":"c91fe577-8e90-411e-b078-8ece235f2f93","Type":"ContainerDied","Data":"b67e23f70f4533511e09fac2edd5683fa3b7f8fdb93aec2eb45cb6e12b33f26b"} Nov 27 17:49:50 crc kubenswrapper[4809]: I1127 17:49:50.892471 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.039773 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nb2g\" (UniqueName: \"kubernetes.io/projected/c91fe577-8e90-411e-b078-8ece235f2f93-kube-api-access-6nb2g\") pod \"c91fe577-8e90-411e-b078-8ece235f2f93\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.040147 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-ssh-key\") pod \"c91fe577-8e90-411e-b078-8ece235f2f93\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.040302 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-inventory\") pod \"c91fe577-8e90-411e-b078-8ece235f2f93\" (UID: \"c91fe577-8e90-411e-b078-8ece235f2f93\") " Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.045047 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c91fe577-8e90-411e-b078-8ece235f2f93-kube-api-access-6nb2g" (OuterVolumeSpecName: "kube-api-access-6nb2g") pod "c91fe577-8e90-411e-b078-8ece235f2f93" (UID: "c91fe577-8e90-411e-b078-8ece235f2f93"). InnerVolumeSpecName "kube-api-access-6nb2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.068177 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-inventory" (OuterVolumeSpecName: "inventory") pod "c91fe577-8e90-411e-b078-8ece235f2f93" (UID: "c91fe577-8e90-411e-b078-8ece235f2f93"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.068530 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c91fe577-8e90-411e-b078-8ece235f2f93" (UID: "c91fe577-8e90-411e-b078-8ece235f2f93"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.142519 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.142711 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91fe577-8e90-411e-b078-8ece235f2f93-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.142815 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nb2g\" (UniqueName: \"kubernetes.io/projected/c91fe577-8e90-411e-b078-8ece235f2f93-kube-api-access-6nb2g\") on node \"crc\" DevicePath \"\"" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.496953 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" event={"ID":"c91fe577-8e90-411e-b078-8ece235f2f93","Type":"ContainerDied","Data":"780dd44306bd10dab3425feacc951e67b671d17c64904bd8d286fd4247bc1050"} Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.497020 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="780dd44306bd10dab3425feacc951e67b671d17c64904bd8d286fd4247bc1050" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.497028 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.567663 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh"] Nov 27 17:49:51 crc kubenswrapper[4809]: E1127 17:49:51.568510 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91fe577-8e90-411e-b078-8ece235f2f93" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.568536 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91fe577-8e90-411e-b078-8ece235f2f93" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.568828 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c91fe577-8e90-411e-b078-8ece235f2f93" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.569643 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.571482 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.571586 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.572226 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.575095 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.581587 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh"] Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.652583 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8wfrh\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.652853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4slz\" (UniqueName: \"kubernetes.io/projected/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-kube-api-access-f4slz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8wfrh\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.653124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8wfrh\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.755017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8wfrh\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.755444 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8wfrh\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.755612 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4slz\" (UniqueName: \"kubernetes.io/projected/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-kube-api-access-f4slz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8wfrh\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.758994 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8wfrh\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.759351 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8wfrh\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.771957 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4slz\" (UniqueName: \"kubernetes.io/projected/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-kube-api-access-f4slz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8wfrh\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:51 crc kubenswrapper[4809]: I1127 17:49:51.886702 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:49:52 crc kubenswrapper[4809]: I1127 17:49:52.511696 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh"] Nov 27 17:49:53 crc kubenswrapper[4809]: I1127 17:49:53.518496 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" event={"ID":"faa8dfb6-0171-4ad3-a68f-46ad6dd83646","Type":"ContainerStarted","Data":"90ccaff3156f3d132eba2102b3653f4e33e48c7f1c5383a41dee919b75657427"} Nov 27 17:49:53 crc kubenswrapper[4809]: I1127 17:49:53.518861 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" event={"ID":"faa8dfb6-0171-4ad3-a68f-46ad6dd83646","Type":"ContainerStarted","Data":"eaa55824b675de0d83326352affaae14f46750588a7e874fc0c8cbe620d9d7c2"} Nov 27 17:49:53 crc kubenswrapper[4809]: I1127 17:49:53.540594 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" podStartSLOduration=2.139112792 podStartE2EDuration="2.540575426s" podCreationTimestamp="2025-11-27 17:49:51 +0000 UTC" firstStartedPulling="2025-11-27 17:49:52.513861045 +0000 UTC m=+2427.786318397" lastFinishedPulling="2025-11-27 17:49:52.915323679 +0000 UTC m=+2428.187781031" observedRunningTime="2025-11-27 17:49:53.539478407 +0000 UTC m=+2428.811935759" watchObservedRunningTime="2025-11-27 17:49:53.540575426 +0000 UTC m=+2428.813032778" Nov 27 17:50:00 crc kubenswrapper[4809]: I1127 17:50:00.458541 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:50:00 crc kubenswrapper[4809]: E1127 17:50:00.459476 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.063710 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-gkh6w"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.085365 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-846b-account-create-update-89v96"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.096349 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-965d-account-create-update-qdw2l"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.106682 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-7b4rl"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.118247 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-965d-account-create-update-qdw2l"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.127371 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-846b-account-create-update-89v96"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.135286 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-r6wt2"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.142796 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-gkh6w"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.150575 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5c84-account-create-update-crb4h"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.166808 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-7b4rl"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.177511 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5c84-account-create-update-crb4h"] Nov 27 17:50:10 crc kubenswrapper[4809]: I1127 17:50:10.187395 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-r6wt2"] Nov 27 17:50:11 crc kubenswrapper[4809]: I1127 17:50:11.473266 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f2a7dfd-6adf-42c5-956d-4f148cea25b2" path="/var/lib/kubelet/pods/0f2a7dfd-6adf-42c5-956d-4f148cea25b2/volumes" Nov 27 17:50:11 crc kubenswrapper[4809]: I1127 17:50:11.475057 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="263114c5-20f2-45c1-b721-2b4d524da56f" path="/var/lib/kubelet/pods/263114c5-20f2-45c1-b721-2b4d524da56f/volumes" Nov 27 17:50:11 crc kubenswrapper[4809]: I1127 17:50:11.476990 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b01b2a4-9229-4e62-8606-44f8ef396563" path="/var/lib/kubelet/pods/4b01b2a4-9229-4e62-8606-44f8ef396563/volumes" Nov 27 17:50:11 crc kubenswrapper[4809]: I1127 17:50:11.478134 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fca3206-aa78-48d2-a774-12ab7e83e2eb" path="/var/lib/kubelet/pods/5fca3206-aa78-48d2-a774-12ab7e83e2eb/volumes" Nov 27 17:50:11 crc kubenswrapper[4809]: I1127 17:50:11.478932 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0435257-8045-4fbb-8cd4-1c2857ec48ae" path="/var/lib/kubelet/pods/a0435257-8045-4fbb-8cd4-1c2857ec48ae/volumes" Nov 27 17:50:11 crc kubenswrapper[4809]: I1127 17:50:11.479822 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f18acc87-ba67-4098-9181-c522e2e22f64" path="/var/lib/kubelet/pods/f18acc87-ba67-4098-9181-c522e2e22f64/volumes" Nov 27 17:50:13 crc kubenswrapper[4809]: I1127 17:50:13.457632 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:50:13 crc kubenswrapper[4809]: E1127 17:50:13.457978 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:50:28 crc kubenswrapper[4809]: I1127 17:50:28.458298 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:50:28 crc kubenswrapper[4809]: E1127 17:50:28.459688 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:50:28 crc kubenswrapper[4809]: I1127 17:50:28.843222 4809 generic.go:334] "Generic (PLEG): container finished" podID="faa8dfb6-0171-4ad3-a68f-46ad6dd83646" containerID="90ccaff3156f3d132eba2102b3653f4e33e48c7f1c5383a41dee919b75657427" exitCode=0 Nov 27 17:50:28 crc kubenswrapper[4809]: I1127 17:50:28.843275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" event={"ID":"faa8dfb6-0171-4ad3-a68f-46ad6dd83646","Type":"ContainerDied","Data":"90ccaff3156f3d132eba2102b3653f4e33e48c7f1c5383a41dee919b75657427"} Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.306240 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.381313 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4slz\" (UniqueName: \"kubernetes.io/projected/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-kube-api-access-f4slz\") pod \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.381564 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-ssh-key\") pod \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.381657 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-inventory\") pod \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\" (UID: \"faa8dfb6-0171-4ad3-a68f-46ad6dd83646\") " Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.387540 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-kube-api-access-f4slz" (OuterVolumeSpecName: "kube-api-access-f4slz") pod "faa8dfb6-0171-4ad3-a68f-46ad6dd83646" (UID: "faa8dfb6-0171-4ad3-a68f-46ad6dd83646"). InnerVolumeSpecName "kube-api-access-f4slz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.412137 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-inventory" (OuterVolumeSpecName: "inventory") pod "faa8dfb6-0171-4ad3-a68f-46ad6dd83646" (UID: "faa8dfb6-0171-4ad3-a68f-46ad6dd83646"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.413493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "faa8dfb6-0171-4ad3-a68f-46ad6dd83646" (UID: "faa8dfb6-0171-4ad3-a68f-46ad6dd83646"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.484214 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.484256 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.484267 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4slz\" (UniqueName: \"kubernetes.io/projected/faa8dfb6-0171-4ad3-a68f-46ad6dd83646-kube-api-access-f4slz\") on node \"crc\" DevicePath \"\"" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.865055 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" event={"ID":"faa8dfb6-0171-4ad3-a68f-46ad6dd83646","Type":"ContainerDied","Data":"eaa55824b675de0d83326352affaae14f46750588a7e874fc0c8cbe620d9d7c2"} Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.865123 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaa55824b675de0d83326352affaae14f46750588a7e874fc0c8cbe620d9d7c2" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.865127 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8wfrh" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.983556 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb"] Nov 27 17:50:30 crc kubenswrapper[4809]: E1127 17:50:30.984117 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faa8dfb6-0171-4ad3-a68f-46ad6dd83646" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.984141 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="faa8dfb6-0171-4ad3-a68f-46ad6dd83646" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.984418 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="faa8dfb6-0171-4ad3-a68f-46ad6dd83646" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.985811 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.990818 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.990834 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.991568 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:50:30 crc kubenswrapper[4809]: I1127 17:50:30.994233 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.004175 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb"] Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.096173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px66k\" (UniqueName: \"kubernetes.io/projected/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-kube-api-access-px66k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.096612 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.096656 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.198497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.198562 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.198705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px66k\" (UniqueName: \"kubernetes.io/projected/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-kube-api-access-px66k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.202612 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.203132 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.218522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px66k\" (UniqueName: \"kubernetes.io/projected/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-kube-api-access-px66k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.303681 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.818168 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb"] Nov 27 17:50:31 crc kubenswrapper[4809]: I1127 17:50:31.876954 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" event={"ID":"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3","Type":"ContainerStarted","Data":"5b020442cafb19af6e107cd24e5708c53f93a0bc526db879b824fd30297237f3"} Nov 27 17:50:32 crc kubenswrapper[4809]: I1127 17:50:32.890310 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" event={"ID":"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3","Type":"ContainerStarted","Data":"b969d8fda775ddde82902ddae36459e09ce42e3791a78d27da4bce798c4d5963"} Nov 27 17:50:32 crc kubenswrapper[4809]: I1127 17:50:32.917185 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" podStartSLOduration=2.264267556 podStartE2EDuration="2.917151546s" podCreationTimestamp="2025-11-27 17:50:30 +0000 UTC" firstStartedPulling="2025-11-27 17:50:31.819473272 +0000 UTC m=+2467.091930624" lastFinishedPulling="2025-11-27 17:50:32.472357262 +0000 UTC m=+2467.744814614" observedRunningTime="2025-11-27 17:50:32.908324695 +0000 UTC m=+2468.180782057" watchObservedRunningTime="2025-11-27 17:50:32.917151546 +0000 UTC m=+2468.189608898" Nov 27 17:50:34 crc kubenswrapper[4809]: I1127 17:50:34.177513 4809 scope.go:117] "RemoveContainer" containerID="6bb89e6b1a9e2e608874ec84a5bf22972f53d03e591dc092dd75ef6f78a63cf6" Nov 27 17:50:34 crc kubenswrapper[4809]: I1127 17:50:34.234828 4809 scope.go:117] "RemoveContainer" containerID="99d1afd96f55db6ed17b3d77809d95307eee5184ad099123dc966d13531c3818" Nov 27 17:50:34 crc kubenswrapper[4809]: I1127 17:50:34.258348 4809 scope.go:117] "RemoveContainer" containerID="9046c1399ae606399c4084ee74581c245bf9bbd1b1bf6aa167c182925ad22071" Nov 27 17:50:34 crc kubenswrapper[4809]: I1127 17:50:34.312904 4809 scope.go:117] "RemoveContainer" containerID="d25e17e852070998e7ac6280459ca4ae23bb3bbdb7b762032229e0d82294fb48" Nov 27 17:50:34 crc kubenswrapper[4809]: I1127 17:50:34.379028 4809 scope.go:117] "RemoveContainer" containerID="d6ed7c84c17e5b83a62f584fc551e0c13e4780b476263cbbf35569276d45f7d3" Nov 27 17:50:34 crc kubenswrapper[4809]: I1127 17:50:34.428152 4809 scope.go:117] "RemoveContainer" containerID="e0b4e8e3098c5cd99e7e226bf1fe4a931793eb0aa9fdfc9ddd33cb32a936029b" Nov 27 17:50:34 crc kubenswrapper[4809]: I1127 17:50:34.501772 4809 scope.go:117] "RemoveContainer" containerID="ad609c13d8ab3eed8a920f1426b5844dd3c60ceffe004309963bbba33ab5ca66" Nov 27 17:50:37 crc kubenswrapper[4809]: I1127 17:50:37.032198 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dnl76"] Nov 27 17:50:37 crc kubenswrapper[4809]: I1127 17:50:37.042009 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dnl76"] Nov 27 17:50:37 crc kubenswrapper[4809]: I1127 17:50:37.470717 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11296c7-060d-409c-852c-509694827f2f" path="/var/lib/kubelet/pods/b11296c7-060d-409c-852c-509694827f2f/volumes" Nov 27 17:50:39 crc kubenswrapper[4809]: I1127 17:50:39.458589 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:50:39 crc kubenswrapper[4809]: E1127 17:50:39.459119 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:50:52 crc kubenswrapper[4809]: I1127 17:50:52.459590 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:50:52 crc kubenswrapper[4809]: E1127 17:50:52.460426 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:50:55 crc kubenswrapper[4809]: I1127 17:50:55.046123 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-wpb8b"] Nov 27 17:50:55 crc kubenswrapper[4809]: I1127 17:50:55.055401 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pks9t"] Nov 27 17:50:55 crc kubenswrapper[4809]: I1127 17:50:55.068000 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-wpb8b"] Nov 27 17:50:55 crc kubenswrapper[4809]: I1127 17:50:55.079062 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pks9t"] Nov 27 17:50:55 crc kubenswrapper[4809]: I1127 17:50:55.475624 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0" path="/var/lib/kubelet/pods/11b3da9f-2b84-4d09-9a10-fc8ffea8f7c0/volumes" Nov 27 17:50:55 crc kubenswrapper[4809]: I1127 17:50:55.477540 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dec0492-5e28-4725-a024-340d7562a2d9" path="/var/lib/kubelet/pods/9dec0492-5e28-4725-a024-340d7562a2d9/volumes" Nov 27 17:51:04 crc kubenswrapper[4809]: I1127 17:51:04.458016 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:51:04 crc kubenswrapper[4809]: E1127 17:51:04.459714 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:51:18 crc kubenswrapper[4809]: I1127 17:51:18.288193 4809 generic.go:334] "Generic (PLEG): container finished" podID="0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3" containerID="b969d8fda775ddde82902ddae36459e09ce42e3791a78d27da4bce798c4d5963" exitCode=0 Nov 27 17:51:18 crc kubenswrapper[4809]: I1127 17:51:18.288287 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" event={"ID":"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3","Type":"ContainerDied","Data":"b969d8fda775ddde82902ddae36459e09ce42e3791a78d27da4bce798c4d5963"} Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.458390 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:51:19 crc kubenswrapper[4809]: E1127 17:51:19.458990 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.700349 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.889212 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-inventory\") pod \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.889480 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px66k\" (UniqueName: \"kubernetes.io/projected/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-kube-api-access-px66k\") pod \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.889572 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-ssh-key\") pod \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\" (UID: \"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3\") " Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.896956 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-kube-api-access-px66k" (OuterVolumeSpecName: "kube-api-access-px66k") pod "0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3" (UID: "0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3"). InnerVolumeSpecName "kube-api-access-px66k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.921527 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3" (UID: "0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.927327 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-inventory" (OuterVolumeSpecName: "inventory") pod "0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3" (UID: "0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.991906 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.991946 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px66k\" (UniqueName: \"kubernetes.io/projected/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-kube-api-access-px66k\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:19 crc kubenswrapper[4809]: I1127 17:51:19.991960 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.305061 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" event={"ID":"0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3","Type":"ContainerDied","Data":"5b020442cafb19af6e107cd24e5708c53f93a0bc526db879b824fd30297237f3"} Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.305114 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b020442cafb19af6e107cd24e5708c53f93a0bc526db879b824fd30297237f3" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.305136 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.392885 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wvz7x"] Nov 27 17:51:20 crc kubenswrapper[4809]: E1127 17:51:20.393802 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.393827 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.394285 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.395811 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.399111 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.399443 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.400008 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.402285 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.404827 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wvz7x\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.404940 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s88b\" (UniqueName: \"kubernetes.io/projected/7616f14b-3549-40b3-ba73-32b1cb830d98-kube-api-access-8s88b\") pod \"ssh-known-hosts-edpm-deployment-wvz7x\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.405104 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wvz7x\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.416906 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wvz7x"] Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.507266 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s88b\" (UniqueName: \"kubernetes.io/projected/7616f14b-3549-40b3-ba73-32b1cb830d98-kube-api-access-8s88b\") pod \"ssh-known-hosts-edpm-deployment-wvz7x\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.507683 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wvz7x\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.507954 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wvz7x\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.521015 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wvz7x\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.524208 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wvz7x\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.525448 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s88b\" (UniqueName: \"kubernetes.io/projected/7616f14b-3549-40b3-ba73-32b1cb830d98-kube-api-access-8s88b\") pod \"ssh-known-hosts-edpm-deployment-wvz7x\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:20 crc kubenswrapper[4809]: I1127 17:51:20.725989 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:21 crc kubenswrapper[4809]: I1127 17:51:21.220070 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wvz7x"] Nov 27 17:51:21 crc kubenswrapper[4809]: I1127 17:51:21.316648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" event={"ID":"7616f14b-3549-40b3-ba73-32b1cb830d98","Type":"ContainerStarted","Data":"18a28224b5da3aa1792a3d1bb4fabb46638d796e5d8cc589567d26ea1caae5de"} Nov 27 17:51:22 crc kubenswrapper[4809]: I1127 17:51:22.327573 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" event={"ID":"7616f14b-3549-40b3-ba73-32b1cb830d98","Type":"ContainerStarted","Data":"7c8d4837dafa6ee9df85e9e6baefd56a1df41af7c3313e5ff0b3299d61c503d1"} Nov 27 17:51:22 crc kubenswrapper[4809]: I1127 17:51:22.345664 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" podStartSLOduration=1.8432650769999999 podStartE2EDuration="2.345647989s" podCreationTimestamp="2025-11-27 17:51:20 +0000 UTC" firstStartedPulling="2025-11-27 17:51:21.225107643 +0000 UTC m=+2516.497564995" lastFinishedPulling="2025-11-27 17:51:21.727490555 +0000 UTC m=+2516.999947907" observedRunningTime="2025-11-27 17:51:22.342955875 +0000 UTC m=+2517.615413227" watchObservedRunningTime="2025-11-27 17:51:22.345647989 +0000 UTC m=+2517.618105341" Nov 27 17:51:28 crc kubenswrapper[4809]: I1127 17:51:28.377297 4809 generic.go:334] "Generic (PLEG): container finished" podID="7616f14b-3549-40b3-ba73-32b1cb830d98" containerID="7c8d4837dafa6ee9df85e9e6baefd56a1df41af7c3313e5ff0b3299d61c503d1" exitCode=0 Nov 27 17:51:28 crc kubenswrapper[4809]: I1127 17:51:28.377416 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" event={"ID":"7616f14b-3549-40b3-ba73-32b1cb830d98","Type":"ContainerDied","Data":"7c8d4837dafa6ee9df85e9e6baefd56a1df41af7c3313e5ff0b3299d61c503d1"} Nov 27 17:51:29 crc kubenswrapper[4809]: I1127 17:51:29.798639 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:29 crc kubenswrapper[4809]: I1127 17:51:29.903447 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-ssh-key-openstack-edpm-ipam\") pod \"7616f14b-3549-40b3-ba73-32b1cb830d98\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " Nov 27 17:51:29 crc kubenswrapper[4809]: I1127 17:51:29.903556 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-inventory-0\") pod \"7616f14b-3549-40b3-ba73-32b1cb830d98\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " Nov 27 17:51:29 crc kubenswrapper[4809]: I1127 17:51:29.903658 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s88b\" (UniqueName: \"kubernetes.io/projected/7616f14b-3549-40b3-ba73-32b1cb830d98-kube-api-access-8s88b\") pod \"7616f14b-3549-40b3-ba73-32b1cb830d98\" (UID: \"7616f14b-3549-40b3-ba73-32b1cb830d98\") " Nov 27 17:51:29 crc kubenswrapper[4809]: I1127 17:51:29.909365 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7616f14b-3549-40b3-ba73-32b1cb830d98-kube-api-access-8s88b" (OuterVolumeSpecName: "kube-api-access-8s88b") pod "7616f14b-3549-40b3-ba73-32b1cb830d98" (UID: "7616f14b-3549-40b3-ba73-32b1cb830d98"). InnerVolumeSpecName "kube-api-access-8s88b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:51:29 crc kubenswrapper[4809]: I1127 17:51:29.931555 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "7616f14b-3549-40b3-ba73-32b1cb830d98" (UID: "7616f14b-3549-40b3-ba73-32b1cb830d98"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:51:29 crc kubenswrapper[4809]: I1127 17:51:29.933128 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7616f14b-3549-40b3-ba73-32b1cb830d98" (UID: "7616f14b-3549-40b3-ba73-32b1cb830d98"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.006708 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.006763 4809 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7616f14b-3549-40b3-ba73-32b1cb830d98-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.006777 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s88b\" (UniqueName: \"kubernetes.io/projected/7616f14b-3549-40b3-ba73-32b1cb830d98-kube-api-access-8s88b\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.394517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" event={"ID":"7616f14b-3549-40b3-ba73-32b1cb830d98","Type":"ContainerDied","Data":"18a28224b5da3aa1792a3d1bb4fabb46638d796e5d8cc589567d26ea1caae5de"} Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.394556 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18a28224b5da3aa1792a3d1bb4fabb46638d796e5d8cc589567d26ea1caae5de" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.394600 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wvz7x" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.473291 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx"] Nov 27 17:51:30 crc kubenswrapper[4809]: E1127 17:51:30.473702 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7616f14b-3549-40b3-ba73-32b1cb830d98" containerName="ssh-known-hosts-edpm-deployment" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.473726 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7616f14b-3549-40b3-ba73-32b1cb830d98" containerName="ssh-known-hosts-edpm-deployment" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.474018 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7616f14b-3549-40b3-ba73-32b1cb830d98" containerName="ssh-known-hosts-edpm-deployment" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.474887 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.477028 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.477050 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.477696 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.479847 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.488312 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx"] Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.617667 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tvrp\" (UniqueName: \"kubernetes.io/projected/5beb9875-497c-4eb8-9b9d-8474a6891b82-kube-api-access-2tvrp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v55jx\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.617954 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v55jx\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.618083 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v55jx\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.720110 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v55jx\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.720265 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v55jx\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.720412 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tvrp\" (UniqueName: \"kubernetes.io/projected/5beb9875-497c-4eb8-9b9d-8474a6891b82-kube-api-access-2tvrp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v55jx\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.723845 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v55jx\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.724042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v55jx\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.737359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tvrp\" (UniqueName: \"kubernetes.io/projected/5beb9875-497c-4eb8-9b9d-8474a6891b82-kube-api-access-2tvrp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v55jx\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:30 crc kubenswrapper[4809]: I1127 17:51:30.794621 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:31 crc kubenswrapper[4809]: I1127 17:51:31.377632 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx"] Nov 27 17:51:31 crc kubenswrapper[4809]: W1127 17:51:31.394458 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5beb9875_497c_4eb8_9b9d_8474a6891b82.slice/crio-08ce95a46512e3b0965efb5c1cb4f408f36b09b07be0d3881e4b75d3077d8e55 WatchSource:0}: Error finding container 08ce95a46512e3b0965efb5c1cb4f408f36b09b07be0d3881e4b75d3077d8e55: Status 404 returned error can't find the container with id 08ce95a46512e3b0965efb5c1cb4f408f36b09b07be0d3881e4b75d3077d8e55 Nov 27 17:51:32 crc kubenswrapper[4809]: I1127 17:51:32.421344 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" event={"ID":"5beb9875-497c-4eb8-9b9d-8474a6891b82","Type":"ContainerStarted","Data":"eeb16df8857f117f679c504be8418b6a587f17af3ec6bb9eac881864d6d57764"} Nov 27 17:51:32 crc kubenswrapper[4809]: I1127 17:51:32.421672 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" event={"ID":"5beb9875-497c-4eb8-9b9d-8474a6891b82","Type":"ContainerStarted","Data":"08ce95a46512e3b0965efb5c1cb4f408f36b09b07be0d3881e4b75d3077d8e55"} Nov 27 17:51:32 crc kubenswrapper[4809]: I1127 17:51:32.444962 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" podStartSLOduration=1.762353086 podStartE2EDuration="2.444940186s" podCreationTimestamp="2025-11-27 17:51:30 +0000 UTC" firstStartedPulling="2025-11-27 17:51:31.405015464 +0000 UTC m=+2526.677472816" lastFinishedPulling="2025-11-27 17:51:32.087602574 +0000 UTC m=+2527.360059916" observedRunningTime="2025-11-27 17:51:32.437120502 +0000 UTC m=+2527.709577854" watchObservedRunningTime="2025-11-27 17:51:32.444940186 +0000 UTC m=+2527.717397538" Nov 27 17:51:32 crc kubenswrapper[4809]: I1127 17:51:32.458608 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:51:32 crc kubenswrapper[4809]: E1127 17:51:32.458935 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:51:34 crc kubenswrapper[4809]: I1127 17:51:34.645390 4809 scope.go:117] "RemoveContainer" containerID="000c8417c4326e932eb3170750207e83f50bbff649ed548350a40a0035a9c1bf" Nov 27 17:51:34 crc kubenswrapper[4809]: I1127 17:51:34.690259 4809 scope.go:117] "RemoveContainer" containerID="6f4fe92641ae06bfe13cdd438b6db28a6cbc7ea22c125760d0fb468229e1a742" Nov 27 17:51:34 crc kubenswrapper[4809]: I1127 17:51:34.742269 4809 scope.go:117] "RemoveContainer" containerID="a6ca040fb04546a5edafdbad849c1f073cbe27d815f8e9bf3146650a709d5ad5" Nov 27 17:51:39 crc kubenswrapper[4809]: I1127 17:51:39.500382 4809 generic.go:334] "Generic (PLEG): container finished" podID="5beb9875-497c-4eb8-9b9d-8474a6891b82" containerID="eeb16df8857f117f679c504be8418b6a587f17af3ec6bb9eac881864d6d57764" exitCode=0 Nov 27 17:51:39 crc kubenswrapper[4809]: I1127 17:51:39.500433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" event={"ID":"5beb9875-497c-4eb8-9b9d-8474a6891b82","Type":"ContainerDied","Data":"eeb16df8857f117f679c504be8418b6a587f17af3ec6bb9eac881864d6d57764"} Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.041176 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-rbmz5"] Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.050696 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-rbmz5"] Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.081483 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.239429 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tvrp\" (UniqueName: \"kubernetes.io/projected/5beb9875-497c-4eb8-9b9d-8474a6891b82-kube-api-access-2tvrp\") pod \"5beb9875-497c-4eb8-9b9d-8474a6891b82\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.240287 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-inventory\") pod \"5beb9875-497c-4eb8-9b9d-8474a6891b82\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.240318 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-ssh-key\") pod \"5beb9875-497c-4eb8-9b9d-8474a6891b82\" (UID: \"5beb9875-497c-4eb8-9b9d-8474a6891b82\") " Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.262049 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5beb9875-497c-4eb8-9b9d-8474a6891b82-kube-api-access-2tvrp" (OuterVolumeSpecName: "kube-api-access-2tvrp") pod "5beb9875-497c-4eb8-9b9d-8474a6891b82" (UID: "5beb9875-497c-4eb8-9b9d-8474a6891b82"). InnerVolumeSpecName "kube-api-access-2tvrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.271215 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5beb9875-497c-4eb8-9b9d-8474a6891b82" (UID: "5beb9875-497c-4eb8-9b9d-8474a6891b82"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.273447 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-inventory" (OuterVolumeSpecName: "inventory") pod "5beb9875-497c-4eb8-9b9d-8474a6891b82" (UID: "5beb9875-497c-4eb8-9b9d-8474a6891b82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.342662 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.342701 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5beb9875-497c-4eb8-9b9d-8474a6891b82-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.342712 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tvrp\" (UniqueName: \"kubernetes.io/projected/5beb9875-497c-4eb8-9b9d-8474a6891b82-kube-api-access-2tvrp\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.471024 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79ca60f5-37e1-4870-8282-7bcb50174a9e" path="/var/lib/kubelet/pods/79ca60f5-37e1-4870-8282-7bcb50174a9e/volumes" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.522622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" event={"ID":"5beb9875-497c-4eb8-9b9d-8474a6891b82","Type":"ContainerDied","Data":"08ce95a46512e3b0965efb5c1cb4f408f36b09b07be0d3881e4b75d3077d8e55"} Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.522674 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v55jx" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.522681 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08ce95a46512e3b0965efb5c1cb4f408f36b09b07be0d3881e4b75d3077d8e55" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.589855 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn"] Nov 27 17:51:41 crc kubenswrapper[4809]: E1127 17:51:41.590498 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5beb9875-497c-4eb8-9b9d-8474a6891b82" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.590525 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5beb9875-497c-4eb8-9b9d-8474a6891b82" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.590825 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5beb9875-497c-4eb8-9b9d-8474a6891b82" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.592041 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.595519 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.595725 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.597256 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.598219 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.604173 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn"] Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.752891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxr89\" (UniqueName: \"kubernetes.io/projected/562a0647-54b3-4b56-9717-24f2ab4ca87f-kube-api-access-pxr89\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.752972 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.753038 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.854873 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxr89\" (UniqueName: \"kubernetes.io/projected/562a0647-54b3-4b56-9717-24f2ab4ca87f-kube-api-access-pxr89\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.854948 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.855010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.862166 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.862649 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.874709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxr89\" (UniqueName: \"kubernetes.io/projected/562a0647-54b3-4b56-9717-24f2ab4ca87f-kube-api-access-pxr89\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:41 crc kubenswrapper[4809]: I1127 17:51:41.914785 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:42 crc kubenswrapper[4809]: I1127 17:51:42.427053 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn"] Nov 27 17:51:42 crc kubenswrapper[4809]: I1127 17:51:42.531560 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" event={"ID":"562a0647-54b3-4b56-9717-24f2ab4ca87f","Type":"ContainerStarted","Data":"8db515577d89df49fa48fdadedd6533b42fea0d4bfc924a19fca0cde02f3af43"} Nov 27 17:51:43 crc kubenswrapper[4809]: I1127 17:51:43.543422 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" event={"ID":"562a0647-54b3-4b56-9717-24f2ab4ca87f","Type":"ContainerStarted","Data":"6d5cdb473e541ad9dd7d16e7482db72ac7f03f2b771120c33ac9f4d4d719d42c"} Nov 27 17:51:43 crc kubenswrapper[4809]: I1127 17:51:43.564807 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" podStartSLOduration=2.026120933 podStartE2EDuration="2.564783733s" podCreationTimestamp="2025-11-27 17:51:41 +0000 UTC" firstStartedPulling="2025-11-27 17:51:42.431602073 +0000 UTC m=+2537.704059425" lastFinishedPulling="2025-11-27 17:51:42.970264873 +0000 UTC m=+2538.242722225" observedRunningTime="2025-11-27 17:51:43.562407219 +0000 UTC m=+2538.834864571" watchObservedRunningTime="2025-11-27 17:51:43.564783733 +0000 UTC m=+2538.837241095" Nov 27 17:51:47 crc kubenswrapper[4809]: I1127 17:51:47.458499 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:51:47 crc kubenswrapper[4809]: E1127 17:51:47.459430 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:51:52 crc kubenswrapper[4809]: I1127 17:51:52.621237 4809 generic.go:334] "Generic (PLEG): container finished" podID="562a0647-54b3-4b56-9717-24f2ab4ca87f" containerID="6d5cdb473e541ad9dd7d16e7482db72ac7f03f2b771120c33ac9f4d4d719d42c" exitCode=0 Nov 27 17:51:52 crc kubenswrapper[4809]: I1127 17:51:52.621332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" event={"ID":"562a0647-54b3-4b56-9717-24f2ab4ca87f","Type":"ContainerDied","Data":"6d5cdb473e541ad9dd7d16e7482db72ac7f03f2b771120c33ac9f4d4d719d42c"} Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.054887 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.195376 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxr89\" (UniqueName: \"kubernetes.io/projected/562a0647-54b3-4b56-9717-24f2ab4ca87f-kube-api-access-pxr89\") pod \"562a0647-54b3-4b56-9717-24f2ab4ca87f\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.195469 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-inventory\") pod \"562a0647-54b3-4b56-9717-24f2ab4ca87f\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.195568 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-ssh-key\") pod \"562a0647-54b3-4b56-9717-24f2ab4ca87f\" (UID: \"562a0647-54b3-4b56-9717-24f2ab4ca87f\") " Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.201640 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/562a0647-54b3-4b56-9717-24f2ab4ca87f-kube-api-access-pxr89" (OuterVolumeSpecName: "kube-api-access-pxr89") pod "562a0647-54b3-4b56-9717-24f2ab4ca87f" (UID: "562a0647-54b3-4b56-9717-24f2ab4ca87f"). InnerVolumeSpecName "kube-api-access-pxr89". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.223051 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-inventory" (OuterVolumeSpecName: "inventory") pod "562a0647-54b3-4b56-9717-24f2ab4ca87f" (UID: "562a0647-54b3-4b56-9717-24f2ab4ca87f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.226049 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "562a0647-54b3-4b56-9717-24f2ab4ca87f" (UID: "562a0647-54b3-4b56-9717-24f2ab4ca87f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.298609 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxr89\" (UniqueName: \"kubernetes.io/projected/562a0647-54b3-4b56-9717-24f2ab4ca87f-kube-api-access-pxr89\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.298643 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.298655 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/562a0647-54b3-4b56-9717-24f2ab4ca87f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.640824 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" event={"ID":"562a0647-54b3-4b56-9717-24f2ab4ca87f","Type":"ContainerDied","Data":"8db515577d89df49fa48fdadedd6533b42fea0d4bfc924a19fca0cde02f3af43"} Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.640866 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8db515577d89df49fa48fdadedd6533b42fea0d4bfc924a19fca0cde02f3af43" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.640893 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.718506 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr"] Nov 27 17:51:54 crc kubenswrapper[4809]: E1127 17:51:54.719019 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562a0647-54b3-4b56-9717-24f2ab4ca87f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.719039 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="562a0647-54b3-4b56-9717-24f2ab4ca87f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.719218 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="562a0647-54b3-4b56-9717-24f2ab4ca87f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.720057 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.723492 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.723818 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.724107 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.724165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.724122 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.724267 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.724596 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.727345 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.729969 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr"] Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.808182 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.808525 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.808719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9smv\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-kube-api-access-w9smv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.808867 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.809035 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.809107 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.809153 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.809344 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.809450 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.809527 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.809619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.809797 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.809970 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.810081 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.911798 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.911840 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.911877 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.911898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.911921 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9smv\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-kube-api-access-w9smv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.911950 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.912000 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.912031 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.912054 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.912102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.912126 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.912142 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.912162 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.912186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.916549 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.916844 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.917157 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.917207 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.917586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.918205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.918567 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.918885 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.919013 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.919430 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.919620 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.919708 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.922381 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:54 crc kubenswrapper[4809]: I1127 17:51:54.937282 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9smv\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-kube-api-access-w9smv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:55 crc kubenswrapper[4809]: I1127 17:51:55.048989 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:51:55 crc kubenswrapper[4809]: I1127 17:51:55.561440 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr"] Nov 27 17:51:55 crc kubenswrapper[4809]: I1127 17:51:55.652360 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" event={"ID":"43c7fb64-1bd5-4341-81e9-6129b4121106","Type":"ContainerStarted","Data":"41e0705f9ef87567ad8f5a5c593383c5e6ba1fa42cbe3d3b11eded37997dcb0e"} Nov 27 17:51:56 crc kubenswrapper[4809]: I1127 17:51:56.667102 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" event={"ID":"43c7fb64-1bd5-4341-81e9-6129b4121106","Type":"ContainerStarted","Data":"c472b2fb89fdaa8dce77ce0432cd7f05e857e4d910389f8b10e6ebddb2241ea8"} Nov 27 17:51:56 crc kubenswrapper[4809]: I1127 17:51:56.692128 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" podStartSLOduration=2.068364957 podStartE2EDuration="2.692105493s" podCreationTimestamp="2025-11-27 17:51:54 +0000 UTC" firstStartedPulling="2025-11-27 17:51:55.55995008 +0000 UTC m=+2550.832407432" lastFinishedPulling="2025-11-27 17:51:56.183690616 +0000 UTC m=+2551.456147968" observedRunningTime="2025-11-27 17:51:56.690965262 +0000 UTC m=+2551.963422614" watchObservedRunningTime="2025-11-27 17:51:56.692105493 +0000 UTC m=+2551.964562855" Nov 27 17:52:02 crc kubenswrapper[4809]: I1127 17:52:02.458009 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:52:02 crc kubenswrapper[4809]: E1127 17:52:02.458841 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:52:17 crc kubenswrapper[4809]: I1127 17:52:17.457930 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:52:17 crc kubenswrapper[4809]: E1127 17:52:17.458756 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:52:30 crc kubenswrapper[4809]: I1127 17:52:30.459350 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:52:30 crc kubenswrapper[4809]: E1127 17:52:30.460235 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:52:31 crc kubenswrapper[4809]: I1127 17:52:31.960029 4809 generic.go:334] "Generic (PLEG): container finished" podID="43c7fb64-1bd5-4341-81e9-6129b4121106" containerID="c472b2fb89fdaa8dce77ce0432cd7f05e857e4d910389f8b10e6ebddb2241ea8" exitCode=0 Nov 27 17:52:31 crc kubenswrapper[4809]: I1127 17:52:31.960117 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" event={"ID":"43c7fb64-1bd5-4341-81e9-6129b4121106","Type":"ContainerDied","Data":"c472b2fb89fdaa8dce77ce0432cd7f05e857e4d910389f8b10e6ebddb2241ea8"} Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.420008 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.527674 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-nova-combined-ca-bundle\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.527899 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-libvirt-combined-ca-bundle\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.527968 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ssh-key\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528053 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528171 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528222 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9smv\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-kube-api-access-w9smv\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528239 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ovn-combined-ca-bundle\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528258 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-telemetry-combined-ca-bundle\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528277 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-repo-setup-combined-ca-bundle\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528296 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-ovn-default-certs-0\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528315 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-neutron-metadata-combined-ca-bundle\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528369 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-bootstrap-combined-ca-bundle\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528449 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.528467 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-inventory\") pod \"43c7fb64-1bd5-4341-81e9-6129b4121106\" (UID: \"43c7fb64-1bd5-4341-81e9-6129b4121106\") " Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.534763 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.535661 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.535854 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.536277 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.536585 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.539271 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.539382 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.540927 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.541048 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-kube-api-access-w9smv" (OuterVolumeSpecName: "kube-api-access-w9smv") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "kube-api-access-w9smv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.541123 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.541187 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.541501 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.567175 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.567345 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-inventory" (OuterVolumeSpecName: "inventory") pod "43c7fb64-1bd5-4341-81e9-6129b4121106" (UID: "43c7fb64-1bd5-4341-81e9-6129b4121106"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.631923 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9smv\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-kube-api-access-w9smv\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.631961 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.631971 4809 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.631981 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.631991 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.632002 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.632011 4809 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.632022 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.632034 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.632042 4809 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.632052 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.632062 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c7fb64-1bd5-4341-81e9-6129b4121106-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.632070 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.632079 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/43c7fb64-1bd5-4341-81e9-6129b4121106-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.976896 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" event={"ID":"43c7fb64-1bd5-4341-81e9-6129b4121106","Type":"ContainerDied","Data":"41e0705f9ef87567ad8f5a5c593383c5e6ba1fa42cbe3d3b11eded37997dcb0e"} Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.976944 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41e0705f9ef87567ad8f5a5c593383c5e6ba1fa42cbe3d3b11eded37997dcb0e" Nov 27 17:52:33 crc kubenswrapper[4809]: I1127 17:52:33.977251 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.065090 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4"] Nov 27 17:52:34 crc kubenswrapper[4809]: E1127 17:52:34.065527 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c7fb64-1bd5-4341-81e9-6129b4121106" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.065549 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c7fb64-1bd5-4341-81e9-6129b4121106" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.065785 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c7fb64-1bd5-4341-81e9-6129b4121106" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.066628 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.068953 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.069088 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.069469 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.069820 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.069851 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.078973 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4"] Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.143067 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.143437 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.143569 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.143694 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cnsq\" (UniqueName: \"kubernetes.io/projected/cc517a71-c16f-4144-94d8-36c2878d89d8-kube-api-access-9cnsq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.143934 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cc517a71-c16f-4144-94d8-36c2878d89d8-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.245672 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.245975 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.246053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cnsq\" (UniqueName: \"kubernetes.io/projected/cc517a71-c16f-4144-94d8-36c2878d89d8-kube-api-access-9cnsq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.246187 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cc517a71-c16f-4144-94d8-36c2878d89d8-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.246440 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.247184 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cc517a71-c16f-4144-94d8-36c2878d89d8-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.249090 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.249424 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.249613 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.263219 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cnsq\" (UniqueName: \"kubernetes.io/projected/cc517a71-c16f-4144-94d8-36c2878d89d8-kube-api-access-9cnsq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dwhp4\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.388226 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.875910 4809 scope.go:117] "RemoveContainer" containerID="3df4fcce6640f111dad80a2d5cb0be2f02cd5c4ad1125a7dd6e4e1105e60decc" Nov 27 17:52:34 crc kubenswrapper[4809]: I1127 17:52:34.893484 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4"] Nov 27 17:52:35 crc kubenswrapper[4809]: I1127 17:52:35.006210 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" event={"ID":"cc517a71-c16f-4144-94d8-36c2878d89d8","Type":"ContainerStarted","Data":"56847bd99ce7c10bd7ffceb02a34e38574ddc97366c95c15452842fa39df5f5a"} Nov 27 17:52:36 crc kubenswrapper[4809]: I1127 17:52:36.016228 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" event={"ID":"cc517a71-c16f-4144-94d8-36c2878d89d8","Type":"ContainerStarted","Data":"d4a1beeed3f8f24354efa73f1d9378fa9981166379989bb514e4d140c7287e6b"} Nov 27 17:52:36 crc kubenswrapper[4809]: I1127 17:52:36.034706 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" podStartSLOduration=1.468080083 podStartE2EDuration="2.034681883s" podCreationTimestamp="2025-11-27 17:52:34 +0000 UTC" firstStartedPulling="2025-11-27 17:52:34.92784223 +0000 UTC m=+2590.200299582" lastFinishedPulling="2025-11-27 17:52:35.49444403 +0000 UTC m=+2590.766901382" observedRunningTime="2025-11-27 17:52:36.030153679 +0000 UTC m=+2591.302611031" watchObservedRunningTime="2025-11-27 17:52:36.034681883 +0000 UTC m=+2591.307139245" Nov 27 17:52:45 crc kubenswrapper[4809]: I1127 17:52:45.464682 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:52:45 crc kubenswrapper[4809]: E1127 17:52:45.465517 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:52:59 crc kubenswrapper[4809]: I1127 17:52:59.457913 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:52:59 crc kubenswrapper[4809]: E1127 17:52:59.458698 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:53:13 crc kubenswrapper[4809]: I1127 17:53:13.458004 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:53:13 crc kubenswrapper[4809]: E1127 17:53:13.461132 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:53:24 crc kubenswrapper[4809]: I1127 17:53:24.458466 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:53:24 crc kubenswrapper[4809]: E1127 17:53:24.459191 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:53:36 crc kubenswrapper[4809]: I1127 17:53:36.553098 4809 generic.go:334] "Generic (PLEG): container finished" podID="cc517a71-c16f-4144-94d8-36c2878d89d8" containerID="d4a1beeed3f8f24354efa73f1d9378fa9981166379989bb514e4d140c7287e6b" exitCode=0 Nov 27 17:53:36 crc kubenswrapper[4809]: I1127 17:53:36.553217 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" event={"ID":"cc517a71-c16f-4144-94d8-36c2878d89d8","Type":"ContainerDied","Data":"d4a1beeed3f8f24354efa73f1d9378fa9981166379989bb514e4d140c7287e6b"} Nov 27 17:53:37 crc kubenswrapper[4809]: I1127 17:53:37.458648 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:53:37 crc kubenswrapper[4809]: E1127 17:53:37.458941 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:53:37 crc kubenswrapper[4809]: I1127 17:53:37.985293 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.076508 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ssh-key\") pod \"cc517a71-c16f-4144-94d8-36c2878d89d8\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.076570 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cc517a71-c16f-4144-94d8-36c2878d89d8-ovncontroller-config-0\") pod \"cc517a71-c16f-4144-94d8-36c2878d89d8\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.076638 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ovn-combined-ca-bundle\") pod \"cc517a71-c16f-4144-94d8-36c2878d89d8\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.076661 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-inventory\") pod \"cc517a71-c16f-4144-94d8-36c2878d89d8\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.076715 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cnsq\" (UniqueName: \"kubernetes.io/projected/cc517a71-c16f-4144-94d8-36c2878d89d8-kube-api-access-9cnsq\") pod \"cc517a71-c16f-4144-94d8-36c2878d89d8\" (UID: \"cc517a71-c16f-4144-94d8-36c2878d89d8\") " Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.084156 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc517a71-c16f-4144-94d8-36c2878d89d8-kube-api-access-9cnsq" (OuterVolumeSpecName: "kube-api-access-9cnsq") pod "cc517a71-c16f-4144-94d8-36c2878d89d8" (UID: "cc517a71-c16f-4144-94d8-36c2878d89d8"). InnerVolumeSpecName "kube-api-access-9cnsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.084470 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "cc517a71-c16f-4144-94d8-36c2878d89d8" (UID: "cc517a71-c16f-4144-94d8-36c2878d89d8"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.108078 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc517a71-c16f-4144-94d8-36c2878d89d8-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "cc517a71-c16f-4144-94d8-36c2878d89d8" (UID: "cc517a71-c16f-4144-94d8-36c2878d89d8"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.110859 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cc517a71-c16f-4144-94d8-36c2878d89d8" (UID: "cc517a71-c16f-4144-94d8-36c2878d89d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.112932 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-inventory" (OuterVolumeSpecName: "inventory") pod "cc517a71-c16f-4144-94d8-36c2878d89d8" (UID: "cc517a71-c16f-4144-94d8-36c2878d89d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.180389 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.180458 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.180472 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cnsq\" (UniqueName: \"kubernetes.io/projected/cc517a71-c16f-4144-94d8-36c2878d89d8-kube-api-access-9cnsq\") on node \"crc\" DevicePath \"\"" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.180482 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc517a71-c16f-4144-94d8-36c2878d89d8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.180496 4809 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cc517a71-c16f-4144-94d8-36c2878d89d8-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.574386 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" event={"ID":"cc517a71-c16f-4144-94d8-36c2878d89d8","Type":"ContainerDied","Data":"56847bd99ce7c10bd7ffceb02a34e38574ddc97366c95c15452842fa39df5f5a"} Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.574428 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56847bd99ce7c10bd7ffceb02a34e38574ddc97366c95c15452842fa39df5f5a" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.574687 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dwhp4" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.654522 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f"] Nov 27 17:53:38 crc kubenswrapper[4809]: E1127 17:53:38.654957 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc517a71-c16f-4144-94d8-36c2878d89d8" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.654977 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc517a71-c16f-4144-94d8-36c2878d89d8" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.655190 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc517a71-c16f-4144-94d8-36c2878d89d8" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.656116 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.658695 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.659711 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.659786 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.659715 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.660076 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.660126 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.664541 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f"] Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.793629 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.793713 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzf59\" (UniqueName: \"kubernetes.io/projected/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-kube-api-access-kzf59\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.793801 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.793863 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.793939 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.794299 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.895702 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.895809 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.895846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzf59\" (UniqueName: \"kubernetes.io/projected/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-kube-api-access-kzf59\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.895904 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.895986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.896191 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.900783 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.901426 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.901433 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.901931 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.902621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.912953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzf59\" (UniqueName: \"kubernetes.io/projected/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-kube-api-access-kzf59\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:38 crc kubenswrapper[4809]: I1127 17:53:38.972649 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:53:39 crc kubenswrapper[4809]: I1127 17:53:39.473980 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f"] Nov 27 17:53:39 crc kubenswrapper[4809]: I1127 17:53:39.479256 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 17:53:39 crc kubenswrapper[4809]: I1127 17:53:39.585178 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" event={"ID":"0d3439fb-cf21-47a9-bbf7-39a34a5469dd","Type":"ContainerStarted","Data":"f9b46bf0c1abbb3bc2989ccff33d68d7fd867c3107e67be2fe335d081e4521de"} Nov 27 17:53:40 crc kubenswrapper[4809]: I1127 17:53:40.595668 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" event={"ID":"0d3439fb-cf21-47a9-bbf7-39a34a5469dd","Type":"ContainerStarted","Data":"a9b1c7d81f1b4a1f2420d670ceba48eb813198b6d3ac3718b1415eb831d7cd30"} Nov 27 17:53:40 crc kubenswrapper[4809]: I1127 17:53:40.616570 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" podStartSLOduration=2.106723326 podStartE2EDuration="2.6165327s" podCreationTimestamp="2025-11-27 17:53:38 +0000 UTC" firstStartedPulling="2025-11-27 17:53:39.479055023 +0000 UTC m=+2654.751512375" lastFinishedPulling="2025-11-27 17:53:39.988864397 +0000 UTC m=+2655.261321749" observedRunningTime="2025-11-27 17:53:40.608396539 +0000 UTC m=+2655.880853891" watchObservedRunningTime="2025-11-27 17:53:40.6165327 +0000 UTC m=+2655.888990052" Nov 27 17:53:52 crc kubenswrapper[4809]: I1127 17:53:52.458498 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:53:52 crc kubenswrapper[4809]: E1127 17:53:52.459294 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:54:07 crc kubenswrapper[4809]: I1127 17:54:07.458649 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:54:07 crc kubenswrapper[4809]: E1127 17:54:07.459469 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:54:18 crc kubenswrapper[4809]: I1127 17:54:18.457784 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:54:18 crc kubenswrapper[4809]: E1127 17:54:18.458599 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 17:54:24 crc kubenswrapper[4809]: I1127 17:54:24.983637 4809 generic.go:334] "Generic (PLEG): container finished" podID="0d3439fb-cf21-47a9-bbf7-39a34a5469dd" containerID="a9b1c7d81f1b4a1f2420d670ceba48eb813198b6d3ac3718b1415eb831d7cd30" exitCode=0 Nov 27 17:54:24 crc kubenswrapper[4809]: I1127 17:54:24.983790 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" event={"ID":"0d3439fb-cf21-47a9-bbf7-39a34a5469dd","Type":"ContainerDied","Data":"a9b1c7d81f1b4a1f2420d670ceba48eb813198b6d3ac3718b1415eb831d7cd30"} Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.420181 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.474816 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-metadata-combined-ca-bundle\") pod \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.475782 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.475901 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-nova-metadata-neutron-config-0\") pod \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.475976 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-ssh-key\") pod \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.476134 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-inventory\") pod \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.476259 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzf59\" (UniqueName: \"kubernetes.io/projected/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-kube-api-access-kzf59\") pod \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\" (UID: \"0d3439fb-cf21-47a9-bbf7-39a34a5469dd\") " Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.482935 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0d3439fb-cf21-47a9-bbf7-39a34a5469dd" (UID: "0d3439fb-cf21-47a9-bbf7-39a34a5469dd"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.483363 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-kube-api-access-kzf59" (OuterVolumeSpecName: "kube-api-access-kzf59") pod "0d3439fb-cf21-47a9-bbf7-39a34a5469dd" (UID: "0d3439fb-cf21-47a9-bbf7-39a34a5469dd"). InnerVolumeSpecName "kube-api-access-kzf59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.509825 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "0d3439fb-cf21-47a9-bbf7-39a34a5469dd" (UID: "0d3439fb-cf21-47a9-bbf7-39a34a5469dd"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.510339 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-inventory" (OuterVolumeSpecName: "inventory") pod "0d3439fb-cf21-47a9-bbf7-39a34a5469dd" (UID: "0d3439fb-cf21-47a9-bbf7-39a34a5469dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.512992 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "0d3439fb-cf21-47a9-bbf7-39a34a5469dd" (UID: "0d3439fb-cf21-47a9-bbf7-39a34a5469dd"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.521974 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0d3439fb-cf21-47a9-bbf7-39a34a5469dd" (UID: "0d3439fb-cf21-47a9-bbf7-39a34a5469dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.579584 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.579619 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzf59\" (UniqueName: \"kubernetes.io/projected/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-kube-api-access-kzf59\") on node \"crc\" DevicePath \"\"" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.579630 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.579639 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.579650 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:54:26 crc kubenswrapper[4809]: I1127 17:54:26.579658 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d3439fb-cf21-47a9-bbf7-39a34a5469dd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.002601 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" event={"ID":"0d3439fb-cf21-47a9-bbf7-39a34a5469dd","Type":"ContainerDied","Data":"f9b46bf0c1abbb3bc2989ccff33d68d7fd867c3107e67be2fe335d081e4521de"} Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.002938 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9b46bf0c1abbb3bc2989ccff33d68d7fd867c3107e67be2fe335d081e4521de" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.002646 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.083181 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp"] Nov 27 17:54:27 crc kubenswrapper[4809]: E1127 17:54:27.083697 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3439fb-cf21-47a9-bbf7-39a34a5469dd" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.083718 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3439fb-cf21-47a9-bbf7-39a34a5469dd" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.083989 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d3439fb-cf21-47a9-bbf7-39a34a5469dd" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.084837 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.087630 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.088145 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.088281 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.088722 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.089670 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.094164 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp"] Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.192885 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.193787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4h22\" (UniqueName: \"kubernetes.io/projected/80c21e8c-3128-48ef-91a8-365409103274-kube-api-access-k4h22\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.194071 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.194178 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.194305 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.296617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.296687 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4h22\" (UniqueName: \"kubernetes.io/projected/80c21e8c-3128-48ef-91a8-365409103274-kube-api-access-k4h22\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.296764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.296792 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.296831 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.301704 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.301773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.302621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.302612 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.315862 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4h22\" (UniqueName: \"kubernetes.io/projected/80c21e8c-3128-48ef-91a8-365409103274-kube-api-access-k4h22\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.417933 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:54:27 crc kubenswrapper[4809]: I1127 17:54:27.930633 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp"] Nov 27 17:54:28 crc kubenswrapper[4809]: I1127 17:54:28.014885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" event={"ID":"80c21e8c-3128-48ef-91a8-365409103274","Type":"ContainerStarted","Data":"d20504ab05d8a60dba3d0f471992a9c9c4f2be76c68f6908ee474bb12135dd5e"} Nov 27 17:54:29 crc kubenswrapper[4809]: I1127 17:54:29.024258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" event={"ID":"80c21e8c-3128-48ef-91a8-365409103274","Type":"ContainerStarted","Data":"af5331d85983f5ca8beb62ae5f05a7420c09dee517c17c2a9dc4ecf7c6477910"} Nov 27 17:54:29 crc kubenswrapper[4809]: I1127 17:54:29.041595 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" podStartSLOduration=1.5680902620000001 podStartE2EDuration="2.041580167s" podCreationTimestamp="2025-11-27 17:54:27 +0000 UTC" firstStartedPulling="2025-11-27 17:54:27.939951735 +0000 UTC m=+2703.212409087" lastFinishedPulling="2025-11-27 17:54:28.41344164 +0000 UTC m=+2703.685898992" observedRunningTime="2025-11-27 17:54:29.039014067 +0000 UTC m=+2704.311471419" watchObservedRunningTime="2025-11-27 17:54:29.041580167 +0000 UTC m=+2704.314037519" Nov 27 17:54:32 crc kubenswrapper[4809]: I1127 17:54:32.458852 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:54:33 crc kubenswrapper[4809]: I1127 17:54:33.063759 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"b076d299448271c4dfcc1047bb2cf65bc76068eb6330d466a9636e70144ea920"} Nov 27 17:56:55 crc kubenswrapper[4809]: I1127 17:56:55.780497 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:56:55 crc kubenswrapper[4809]: I1127 17:56:55.781390 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:57:19 crc kubenswrapper[4809]: I1127 17:57:19.986829 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qt8g8"] Nov 27 17:57:19 crc kubenswrapper[4809]: I1127 17:57:19.991354 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.005063 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qt8g8"] Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.058805 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqhsx\" (UniqueName: \"kubernetes.io/projected/9e6fdb03-d99b-44ea-abca-7eefa84fab50-kube-api-access-qqhsx\") pod \"certified-operators-qt8g8\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.058878 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-utilities\") pod \"certified-operators-qt8g8\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.058896 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-catalog-content\") pod \"certified-operators-qt8g8\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.161021 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqhsx\" (UniqueName: \"kubernetes.io/projected/9e6fdb03-d99b-44ea-abca-7eefa84fab50-kube-api-access-qqhsx\") pod \"certified-operators-qt8g8\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.161376 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-utilities\") pod \"certified-operators-qt8g8\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.161515 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-catalog-content\") pod \"certified-operators-qt8g8\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.162029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-catalog-content\") pod \"certified-operators-qt8g8\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.162283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-utilities\") pod \"certified-operators-qt8g8\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.185300 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqhsx\" (UniqueName: \"kubernetes.io/projected/9e6fdb03-d99b-44ea-abca-7eefa84fab50-kube-api-access-qqhsx\") pod \"certified-operators-qt8g8\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.313732 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:20 crc kubenswrapper[4809]: I1127 17:57:20.833413 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qt8g8"] Nov 27 17:57:21 crc kubenswrapper[4809]: I1127 17:57:21.708100 4809 generic.go:334] "Generic (PLEG): container finished" podID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerID="312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937" exitCode=0 Nov 27 17:57:21 crc kubenswrapper[4809]: I1127 17:57:21.708342 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qt8g8" event={"ID":"9e6fdb03-d99b-44ea-abca-7eefa84fab50","Type":"ContainerDied","Data":"312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937"} Nov 27 17:57:21 crc kubenswrapper[4809]: I1127 17:57:21.708629 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qt8g8" event={"ID":"9e6fdb03-d99b-44ea-abca-7eefa84fab50","Type":"ContainerStarted","Data":"513d43c72a21d6f5ee8cbadf4bf3b420105a58b59dea40d255233ecb2574fd67"} Nov 27 17:57:22 crc kubenswrapper[4809]: I1127 17:57:22.718652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qt8g8" event={"ID":"9e6fdb03-d99b-44ea-abca-7eefa84fab50","Type":"ContainerStarted","Data":"b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912"} Nov 27 17:57:23 crc kubenswrapper[4809]: I1127 17:57:23.733098 4809 generic.go:334] "Generic (PLEG): container finished" podID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerID="b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912" exitCode=0 Nov 27 17:57:23 crc kubenswrapper[4809]: I1127 17:57:23.733174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qt8g8" event={"ID":"9e6fdb03-d99b-44ea-abca-7eefa84fab50","Type":"ContainerDied","Data":"b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912"} Nov 27 17:57:24 crc kubenswrapper[4809]: I1127 17:57:24.744068 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qt8g8" event={"ID":"9e6fdb03-d99b-44ea-abca-7eefa84fab50","Type":"ContainerStarted","Data":"875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb"} Nov 27 17:57:24 crc kubenswrapper[4809]: I1127 17:57:24.770858 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qt8g8" podStartSLOduration=3.2281438 podStartE2EDuration="5.770827605s" podCreationTimestamp="2025-11-27 17:57:19 +0000 UTC" firstStartedPulling="2025-11-27 17:57:21.710732459 +0000 UTC m=+2876.983189811" lastFinishedPulling="2025-11-27 17:57:24.253416264 +0000 UTC m=+2879.525873616" observedRunningTime="2025-11-27 17:57:24.760827802 +0000 UTC m=+2880.033285164" watchObservedRunningTime="2025-11-27 17:57:24.770827605 +0000 UTC m=+2880.043284957" Nov 27 17:57:25 crc kubenswrapper[4809]: I1127 17:57:25.779144 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:57:25 crc kubenswrapper[4809]: I1127 17:57:25.779205 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:57:30 crc kubenswrapper[4809]: I1127 17:57:30.314147 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:30 crc kubenswrapper[4809]: I1127 17:57:30.315725 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:30 crc kubenswrapper[4809]: I1127 17:57:30.364807 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:30 crc kubenswrapper[4809]: I1127 17:57:30.845867 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:30 crc kubenswrapper[4809]: I1127 17:57:30.899381 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qt8g8"] Nov 27 17:57:32 crc kubenswrapper[4809]: I1127 17:57:32.814872 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qt8g8" podUID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerName="registry-server" containerID="cri-o://875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb" gracePeriod=2 Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.264695 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.469459 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-utilities\") pod \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.469845 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-catalog-content\") pod \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.469986 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqhsx\" (UniqueName: \"kubernetes.io/projected/9e6fdb03-d99b-44ea-abca-7eefa84fab50-kube-api-access-qqhsx\") pod \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\" (UID: \"9e6fdb03-d99b-44ea-abca-7eefa84fab50\") " Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.470594 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-utilities" (OuterVolumeSpecName: "utilities") pod "9e6fdb03-d99b-44ea-abca-7eefa84fab50" (UID: "9e6fdb03-d99b-44ea-abca-7eefa84fab50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.479972 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6fdb03-d99b-44ea-abca-7eefa84fab50-kube-api-access-qqhsx" (OuterVolumeSpecName: "kube-api-access-qqhsx") pod "9e6fdb03-d99b-44ea-abca-7eefa84fab50" (UID: "9e6fdb03-d99b-44ea-abca-7eefa84fab50"). InnerVolumeSpecName "kube-api-access-qqhsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.574688 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqhsx\" (UniqueName: \"kubernetes.io/projected/9e6fdb03-d99b-44ea-abca-7eefa84fab50-kube-api-access-qqhsx\") on node \"crc\" DevicePath \"\"" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.574774 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.825609 4809 generic.go:334] "Generic (PLEG): container finished" podID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerID="875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb" exitCode=0 Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.825673 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qt8g8" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.825706 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qt8g8" event={"ID":"9e6fdb03-d99b-44ea-abca-7eefa84fab50","Type":"ContainerDied","Data":"875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb"} Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.826841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qt8g8" event={"ID":"9e6fdb03-d99b-44ea-abca-7eefa84fab50","Type":"ContainerDied","Data":"513d43c72a21d6f5ee8cbadf4bf3b420105a58b59dea40d255233ecb2574fd67"} Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.826865 4809 scope.go:117] "RemoveContainer" containerID="875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.849696 4809 scope.go:117] "RemoveContainer" containerID="b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.868967 4809 scope.go:117] "RemoveContainer" containerID="312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.918240 4809 scope.go:117] "RemoveContainer" containerID="875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb" Nov 27 17:57:33 crc kubenswrapper[4809]: E1127 17:57:33.918675 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb\": container with ID starting with 875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb not found: ID does not exist" containerID="875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.918719 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb"} err="failed to get container status \"875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb\": rpc error: code = NotFound desc = could not find container \"875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb\": container with ID starting with 875672cc180b7032e47e32cae0257c1007cd8e709111670c9b39d7e114d372cb not found: ID does not exist" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.918760 4809 scope.go:117] "RemoveContainer" containerID="b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912" Nov 27 17:57:33 crc kubenswrapper[4809]: E1127 17:57:33.919068 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912\": container with ID starting with b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912 not found: ID does not exist" containerID="b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.919106 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912"} err="failed to get container status \"b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912\": rpc error: code = NotFound desc = could not find container \"b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912\": container with ID starting with b11ad56e92d15538777e43a40deea0ef13e37ea2fb25607a94f0e96f659ee912 not found: ID does not exist" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.919126 4809 scope.go:117] "RemoveContainer" containerID="312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937" Nov 27 17:57:33 crc kubenswrapper[4809]: E1127 17:57:33.919386 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937\": container with ID starting with 312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937 not found: ID does not exist" containerID="312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937" Nov 27 17:57:33 crc kubenswrapper[4809]: I1127 17:57:33.919421 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937"} err="failed to get container status \"312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937\": rpc error: code = NotFound desc = could not find container \"312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937\": container with ID starting with 312bd571219eb566ffd795ed43ed8a5b2be7bb5ab11538009d3faa9b4827b937 not found: ID does not exist" Nov 27 17:57:34 crc kubenswrapper[4809]: I1127 17:57:34.046372 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e6fdb03-d99b-44ea-abca-7eefa84fab50" (UID: "9e6fdb03-d99b-44ea-abca-7eefa84fab50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:57:34 crc kubenswrapper[4809]: I1127 17:57:34.086345 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6fdb03-d99b-44ea-abca-7eefa84fab50-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:57:34 crc kubenswrapper[4809]: I1127 17:57:34.163378 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qt8g8"] Nov 27 17:57:34 crc kubenswrapper[4809]: I1127 17:57:34.171962 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qt8g8"] Nov 27 17:57:35 crc kubenswrapper[4809]: I1127 17:57:35.470581 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" path="/var/lib/kubelet/pods/9e6fdb03-d99b-44ea-abca-7eefa84fab50/volumes" Nov 27 17:57:55 crc kubenswrapper[4809]: I1127 17:57:55.779499 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 17:57:55 crc kubenswrapper[4809]: I1127 17:57:55.780104 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 17:57:55 crc kubenswrapper[4809]: I1127 17:57:55.780160 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 17:57:55 crc kubenswrapper[4809]: I1127 17:57:55.781020 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b076d299448271c4dfcc1047bb2cf65bc76068eb6330d466a9636e70144ea920"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 17:57:55 crc kubenswrapper[4809]: I1127 17:57:55.781086 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://b076d299448271c4dfcc1047bb2cf65bc76068eb6330d466a9636e70144ea920" gracePeriod=600 Nov 27 17:57:56 crc kubenswrapper[4809]: I1127 17:57:56.044516 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="b076d299448271c4dfcc1047bb2cf65bc76068eb6330d466a9636e70144ea920" exitCode=0 Nov 27 17:57:56 crc kubenswrapper[4809]: I1127 17:57:56.044554 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"b076d299448271c4dfcc1047bb2cf65bc76068eb6330d466a9636e70144ea920"} Nov 27 17:57:56 crc kubenswrapper[4809]: I1127 17:57:56.044623 4809 scope.go:117] "RemoveContainer" containerID="9052fb4512b444439aa956582ebce589136de82adac7ccfd4dbed38fe17e46b0" Nov 27 17:57:57 crc kubenswrapper[4809]: I1127 17:57:57.055672 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064"} Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.159430 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q9vx5"] Nov 27 17:58:02 crc kubenswrapper[4809]: E1127 17:58:02.161215 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerName="registry-server" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.161231 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerName="registry-server" Nov 27 17:58:02 crc kubenswrapper[4809]: E1127 17:58:02.161247 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerName="extract-utilities" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.161257 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerName="extract-utilities" Nov 27 17:58:02 crc kubenswrapper[4809]: E1127 17:58:02.161268 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerName="extract-content" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.161275 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerName="extract-content" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.161592 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6fdb03-d99b-44ea-abca-7eefa84fab50" containerName="registry-server" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.163984 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.174479 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9vx5"] Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.273976 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-utilities\") pod \"redhat-operators-q9vx5\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.274436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-catalog-content\") pod \"redhat-operators-q9vx5\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.274607 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj894\" (UniqueName: \"kubernetes.io/projected/eae66b3c-224c-47e6-88d2-21fc1e9dd437-kube-api-access-bj894\") pod \"redhat-operators-q9vx5\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.377899 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-catalog-content\") pod \"redhat-operators-q9vx5\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.378019 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj894\" (UniqueName: \"kubernetes.io/projected/eae66b3c-224c-47e6-88d2-21fc1e9dd437-kube-api-access-bj894\") pod \"redhat-operators-q9vx5\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.378232 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-utilities\") pod \"redhat-operators-q9vx5\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.378625 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-catalog-content\") pod \"redhat-operators-q9vx5\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.378970 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-utilities\") pod \"redhat-operators-q9vx5\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.400706 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj894\" (UniqueName: \"kubernetes.io/projected/eae66b3c-224c-47e6-88d2-21fc1e9dd437-kube-api-access-bj894\") pod \"redhat-operators-q9vx5\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.499847 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:02 crc kubenswrapper[4809]: I1127 17:58:02.996126 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9vx5"] Nov 27 17:58:03 crc kubenswrapper[4809]: I1127 17:58:03.122270 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9vx5" event={"ID":"eae66b3c-224c-47e6-88d2-21fc1e9dd437","Type":"ContainerStarted","Data":"38045aa48e76e1f6ad2d6449464097853ecb493843bf3440b6ef7c85650d8985"} Nov 27 17:58:04 crc kubenswrapper[4809]: I1127 17:58:04.134061 4809 generic.go:334] "Generic (PLEG): container finished" podID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerID="42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80" exitCode=0 Nov 27 17:58:04 crc kubenswrapper[4809]: I1127 17:58:04.134112 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9vx5" event={"ID":"eae66b3c-224c-47e6-88d2-21fc1e9dd437","Type":"ContainerDied","Data":"42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80"} Nov 27 17:58:06 crc kubenswrapper[4809]: I1127 17:58:06.152453 4809 generic.go:334] "Generic (PLEG): container finished" podID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerID="35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3" exitCode=0 Nov 27 17:58:06 crc kubenswrapper[4809]: I1127 17:58:06.152530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9vx5" event={"ID":"eae66b3c-224c-47e6-88d2-21fc1e9dd437","Type":"ContainerDied","Data":"35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3"} Nov 27 17:58:08 crc kubenswrapper[4809]: I1127 17:58:08.174432 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9vx5" event={"ID":"eae66b3c-224c-47e6-88d2-21fc1e9dd437","Type":"ContainerStarted","Data":"fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d"} Nov 27 17:58:08 crc kubenswrapper[4809]: I1127 17:58:08.191573 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q9vx5" podStartSLOduration=2.908738186 podStartE2EDuration="6.191553057s" podCreationTimestamp="2025-11-27 17:58:02 +0000 UTC" firstStartedPulling="2025-11-27 17:58:04.136417103 +0000 UTC m=+2919.408874465" lastFinishedPulling="2025-11-27 17:58:07.419231984 +0000 UTC m=+2922.691689336" observedRunningTime="2025-11-27 17:58:08.190137538 +0000 UTC m=+2923.462594900" watchObservedRunningTime="2025-11-27 17:58:08.191553057 +0000 UTC m=+2923.464010419" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.344025 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f724q"] Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.346859 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.359571 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f724q"] Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.470257 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-catalog-content\") pod \"redhat-marketplace-f724q\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.470491 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-utilities\") pod \"redhat-marketplace-f724q\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.470901 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bg4n\" (UniqueName: \"kubernetes.io/projected/a55fb83c-f00e-406d-ba12-9e0dd182f98d-kube-api-access-6bg4n\") pod \"redhat-marketplace-f724q\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.500788 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.500850 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.572520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-catalog-content\") pod \"redhat-marketplace-f724q\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.572665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-utilities\") pod \"redhat-marketplace-f724q\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.572819 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bg4n\" (UniqueName: \"kubernetes.io/projected/a55fb83c-f00e-406d-ba12-9e0dd182f98d-kube-api-access-6bg4n\") pod \"redhat-marketplace-f724q\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.573277 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-utilities\") pod \"redhat-marketplace-f724q\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.573301 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-catalog-content\") pod \"redhat-marketplace-f724q\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.600852 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bg4n\" (UniqueName: \"kubernetes.io/projected/a55fb83c-f00e-406d-ba12-9e0dd182f98d-kube-api-access-6bg4n\") pod \"redhat-marketplace-f724q\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:12 crc kubenswrapper[4809]: I1127 17:58:12.675537 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:13 crc kubenswrapper[4809]: I1127 17:58:13.142132 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f724q"] Nov 27 17:58:13 crc kubenswrapper[4809]: W1127 17:58:13.145298 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda55fb83c_f00e_406d_ba12_9e0dd182f98d.slice/crio-699bd142f37a37bb8034dafa655cf70c686b0895a1f081c6699718cf72cb1f00 WatchSource:0}: Error finding container 699bd142f37a37bb8034dafa655cf70c686b0895a1f081c6699718cf72cb1f00: Status 404 returned error can't find the container with id 699bd142f37a37bb8034dafa655cf70c686b0895a1f081c6699718cf72cb1f00 Nov 27 17:58:13 crc kubenswrapper[4809]: I1127 17:58:13.221550 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f724q" event={"ID":"a55fb83c-f00e-406d-ba12-9e0dd182f98d","Type":"ContainerStarted","Data":"699bd142f37a37bb8034dafa655cf70c686b0895a1f081c6699718cf72cb1f00"} Nov 27 17:58:13 crc kubenswrapper[4809]: I1127 17:58:13.546592 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q9vx5" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerName="registry-server" probeResult="failure" output=< Nov 27 17:58:13 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 27 17:58:13 crc kubenswrapper[4809]: > Nov 27 17:58:14 crc kubenswrapper[4809]: I1127 17:58:14.232102 4809 generic.go:334] "Generic (PLEG): container finished" podID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerID="faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f" exitCode=0 Nov 27 17:58:14 crc kubenswrapper[4809]: I1127 17:58:14.232708 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f724q" event={"ID":"a55fb83c-f00e-406d-ba12-9e0dd182f98d","Type":"ContainerDied","Data":"faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f"} Nov 27 17:58:15 crc kubenswrapper[4809]: I1127 17:58:15.247341 4809 generic.go:334] "Generic (PLEG): container finished" podID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerID="025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773" exitCode=0 Nov 27 17:58:15 crc kubenswrapper[4809]: I1127 17:58:15.247847 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f724q" event={"ID":"a55fb83c-f00e-406d-ba12-9e0dd182f98d","Type":"ContainerDied","Data":"025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773"} Nov 27 17:58:16 crc kubenswrapper[4809]: I1127 17:58:16.259494 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f724q" event={"ID":"a55fb83c-f00e-406d-ba12-9e0dd182f98d","Type":"ContainerStarted","Data":"2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69"} Nov 27 17:58:16 crc kubenswrapper[4809]: I1127 17:58:16.282633 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f724q" podStartSLOduration=2.779611181 podStartE2EDuration="4.282613642s" podCreationTimestamp="2025-11-27 17:58:12 +0000 UTC" firstStartedPulling="2025-11-27 17:58:14.234713981 +0000 UTC m=+2929.507171333" lastFinishedPulling="2025-11-27 17:58:15.737716442 +0000 UTC m=+2931.010173794" observedRunningTime="2025-11-27 17:58:16.279561609 +0000 UTC m=+2931.552018971" watchObservedRunningTime="2025-11-27 17:58:16.282613642 +0000 UTC m=+2931.555070994" Nov 27 17:58:22 crc kubenswrapper[4809]: I1127 17:58:22.610385 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:22 crc kubenswrapper[4809]: I1127 17:58:22.662903 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:22 crc kubenswrapper[4809]: I1127 17:58:22.675829 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:22 crc kubenswrapper[4809]: I1127 17:58:22.675875 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:22 crc kubenswrapper[4809]: I1127 17:58:22.725310 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:22 crc kubenswrapper[4809]: I1127 17:58:22.849674 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q9vx5"] Nov 27 17:58:23 crc kubenswrapper[4809]: I1127 17:58:23.371233 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:24 crc kubenswrapper[4809]: I1127 17:58:24.333573 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q9vx5" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerName="registry-server" containerID="cri-o://fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d" gracePeriod=2 Nov 27 17:58:24 crc kubenswrapper[4809]: I1127 17:58:24.823290 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:24 crc kubenswrapper[4809]: I1127 17:58:24.930989 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-utilities\") pod \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " Nov 27 17:58:24 crc kubenswrapper[4809]: I1127 17:58:24.931077 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj894\" (UniqueName: \"kubernetes.io/projected/eae66b3c-224c-47e6-88d2-21fc1e9dd437-kube-api-access-bj894\") pod \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " Nov 27 17:58:24 crc kubenswrapper[4809]: I1127 17:58:24.931298 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-catalog-content\") pod \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\" (UID: \"eae66b3c-224c-47e6-88d2-21fc1e9dd437\") " Nov 27 17:58:24 crc kubenswrapper[4809]: I1127 17:58:24.932487 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-utilities" (OuterVolumeSpecName: "utilities") pod "eae66b3c-224c-47e6-88d2-21fc1e9dd437" (UID: "eae66b3c-224c-47e6-88d2-21fc1e9dd437"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:58:24 crc kubenswrapper[4809]: I1127 17:58:24.937778 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae66b3c-224c-47e6-88d2-21fc1e9dd437-kube-api-access-bj894" (OuterVolumeSpecName: "kube-api-access-bj894") pod "eae66b3c-224c-47e6-88d2-21fc1e9dd437" (UID: "eae66b3c-224c-47e6-88d2-21fc1e9dd437"). InnerVolumeSpecName "kube-api-access-bj894". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.034653 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.034692 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj894\" (UniqueName: \"kubernetes.io/projected/eae66b3c-224c-47e6-88d2-21fc1e9dd437-kube-api-access-bj894\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.042919 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eae66b3c-224c-47e6-88d2-21fc1e9dd437" (UID: "eae66b3c-224c-47e6-88d2-21fc1e9dd437"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.054278 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f724q"] Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.136830 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae66b3c-224c-47e6-88d2-21fc1e9dd437-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.343778 4809 generic.go:334] "Generic (PLEG): container finished" podID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerID="fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d" exitCode=0 Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.343960 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f724q" podUID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerName="registry-server" containerID="cri-o://2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69" gracePeriod=2 Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.344221 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9vx5" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.345521 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9vx5" event={"ID":"eae66b3c-224c-47e6-88d2-21fc1e9dd437","Type":"ContainerDied","Data":"fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d"} Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.345575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9vx5" event={"ID":"eae66b3c-224c-47e6-88d2-21fc1e9dd437","Type":"ContainerDied","Data":"38045aa48e76e1f6ad2d6449464097853ecb493843bf3440b6ef7c85650d8985"} Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.345594 4809 scope.go:117] "RemoveContainer" containerID="fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.367507 4809 scope.go:117] "RemoveContainer" containerID="35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.382325 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q9vx5"] Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.400764 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q9vx5"] Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.413288 4809 scope.go:117] "RemoveContainer" containerID="42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.479077 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" path="/var/lib/kubelet/pods/eae66b3c-224c-47e6-88d2-21fc1e9dd437/volumes" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.546331 4809 scope.go:117] "RemoveContainer" containerID="fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d" Nov 27 17:58:25 crc kubenswrapper[4809]: E1127 17:58:25.547141 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d\": container with ID starting with fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d not found: ID does not exist" containerID="fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.547180 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d"} err="failed to get container status \"fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d\": rpc error: code = NotFound desc = could not find container \"fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d\": container with ID starting with fb7ee7b826f273d4b2d7dbb60dc68bb909e3e89653fd7097c8394a22adc0eb9d not found: ID does not exist" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.547226 4809 scope.go:117] "RemoveContainer" containerID="35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3" Nov 27 17:58:25 crc kubenswrapper[4809]: E1127 17:58:25.547597 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3\": container with ID starting with 35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3 not found: ID does not exist" containerID="35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.547635 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3"} err="failed to get container status \"35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3\": rpc error: code = NotFound desc = could not find container \"35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3\": container with ID starting with 35eb9e87908e813e1cfd25a476e24ecd6847b078dbeefebec118c16f2ec880e3 not found: ID does not exist" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.547677 4809 scope.go:117] "RemoveContainer" containerID="42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80" Nov 27 17:58:25 crc kubenswrapper[4809]: E1127 17:58:25.548058 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80\": container with ID starting with 42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80 not found: ID does not exist" containerID="42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.548082 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80"} err="failed to get container status \"42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80\": rpc error: code = NotFound desc = could not find container \"42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80\": container with ID starting with 42371df8819e5e334c0bc6f7934b690bc6dfe2097ee9a581a787185cf1fe1f80 not found: ID does not exist" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.862280 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.950674 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bg4n\" (UniqueName: \"kubernetes.io/projected/a55fb83c-f00e-406d-ba12-9e0dd182f98d-kube-api-access-6bg4n\") pod \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.950911 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-catalog-content\") pod \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.950979 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-utilities\") pod \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\" (UID: \"a55fb83c-f00e-406d-ba12-9e0dd182f98d\") " Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.951723 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-utilities" (OuterVolumeSpecName: "utilities") pod "a55fb83c-f00e-406d-ba12-9e0dd182f98d" (UID: "a55fb83c-f00e-406d-ba12-9e0dd182f98d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.956825 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a55fb83c-f00e-406d-ba12-9e0dd182f98d-kube-api-access-6bg4n" (OuterVolumeSpecName: "kube-api-access-6bg4n") pod "a55fb83c-f00e-406d-ba12-9e0dd182f98d" (UID: "a55fb83c-f00e-406d-ba12-9e0dd182f98d"). InnerVolumeSpecName "kube-api-access-6bg4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:58:25 crc kubenswrapper[4809]: I1127 17:58:25.971781 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a55fb83c-f00e-406d-ba12-9e0dd182f98d" (UID: "a55fb83c-f00e-406d-ba12-9e0dd182f98d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.053622 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bg4n\" (UniqueName: \"kubernetes.io/projected/a55fb83c-f00e-406d-ba12-9e0dd182f98d-kube-api-access-6bg4n\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.053666 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.053678 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55fb83c-f00e-406d-ba12-9e0dd182f98d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.359388 4809 generic.go:334] "Generic (PLEG): container finished" podID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerID="2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69" exitCode=0 Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.359445 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f724q" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.359445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f724q" event={"ID":"a55fb83c-f00e-406d-ba12-9e0dd182f98d","Type":"ContainerDied","Data":"2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69"} Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.359593 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f724q" event={"ID":"a55fb83c-f00e-406d-ba12-9e0dd182f98d","Type":"ContainerDied","Data":"699bd142f37a37bb8034dafa655cf70c686b0895a1f081c6699718cf72cb1f00"} Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.359619 4809 scope.go:117] "RemoveContainer" containerID="2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.361617 4809 generic.go:334] "Generic (PLEG): container finished" podID="80c21e8c-3128-48ef-91a8-365409103274" containerID="af5331d85983f5ca8beb62ae5f05a7420c09dee517c17c2a9dc4ecf7c6477910" exitCode=0 Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.361649 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" event={"ID":"80c21e8c-3128-48ef-91a8-365409103274","Type":"ContainerDied","Data":"af5331d85983f5ca8beb62ae5f05a7420c09dee517c17c2a9dc4ecf7c6477910"} Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.400990 4809 scope.go:117] "RemoveContainer" containerID="025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.409665 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f724q"] Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.417611 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f724q"] Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.445769 4809 scope.go:117] "RemoveContainer" containerID="faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.462785 4809 scope.go:117] "RemoveContainer" containerID="2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69" Nov 27 17:58:26 crc kubenswrapper[4809]: E1127 17:58:26.463296 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69\": container with ID starting with 2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69 not found: ID does not exist" containerID="2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.463347 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69"} err="failed to get container status \"2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69\": rpc error: code = NotFound desc = could not find container \"2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69\": container with ID starting with 2b33625287cec0fa9bb65b9d74588851f787061fd0984da72c3e12c49cb54d69 not found: ID does not exist" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.463379 4809 scope.go:117] "RemoveContainer" containerID="025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773" Nov 27 17:58:26 crc kubenswrapper[4809]: E1127 17:58:26.463820 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773\": container with ID starting with 025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773 not found: ID does not exist" containerID="025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.463849 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773"} err="failed to get container status \"025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773\": rpc error: code = NotFound desc = could not find container \"025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773\": container with ID starting with 025584e07fdafb6974439387667ff2540914cef3a195bf6bfb8e016705fb0773 not found: ID does not exist" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.463871 4809 scope.go:117] "RemoveContainer" containerID="faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f" Nov 27 17:58:26 crc kubenswrapper[4809]: E1127 17:58:26.464124 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f\": container with ID starting with faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f not found: ID does not exist" containerID="faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f" Nov 27 17:58:26 crc kubenswrapper[4809]: I1127 17:58:26.464170 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f"} err="failed to get container status \"faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f\": rpc error: code = NotFound desc = could not find container \"faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f\": container with ID starting with faee791a8ca8fb87867169c14e6ba3e58cd62c989d26668e1b9338e791ab1e3f not found: ID does not exist" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.484163 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" path="/var/lib/kubelet/pods/a55fb83c-f00e-406d-ba12-9e0dd182f98d/volumes" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.786637 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.889402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-inventory\") pod \"80c21e8c-3128-48ef-91a8-365409103274\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.889461 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-combined-ca-bundle\") pod \"80c21e8c-3128-48ef-91a8-365409103274\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.889692 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-secret-0\") pod \"80c21e8c-3128-48ef-91a8-365409103274\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.889834 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4h22\" (UniqueName: \"kubernetes.io/projected/80c21e8c-3128-48ef-91a8-365409103274-kube-api-access-k4h22\") pod \"80c21e8c-3128-48ef-91a8-365409103274\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.889893 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-ssh-key\") pod \"80c21e8c-3128-48ef-91a8-365409103274\" (UID: \"80c21e8c-3128-48ef-91a8-365409103274\") " Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.895017 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "80c21e8c-3128-48ef-91a8-365409103274" (UID: "80c21e8c-3128-48ef-91a8-365409103274"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.895248 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80c21e8c-3128-48ef-91a8-365409103274-kube-api-access-k4h22" (OuterVolumeSpecName: "kube-api-access-k4h22") pod "80c21e8c-3128-48ef-91a8-365409103274" (UID: "80c21e8c-3128-48ef-91a8-365409103274"). InnerVolumeSpecName "kube-api-access-k4h22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.917775 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "80c21e8c-3128-48ef-91a8-365409103274" (UID: "80c21e8c-3128-48ef-91a8-365409103274"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.919057 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-inventory" (OuterVolumeSpecName: "inventory") pod "80c21e8c-3128-48ef-91a8-365409103274" (UID: "80c21e8c-3128-48ef-91a8-365409103274"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.925282 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "80c21e8c-3128-48ef-91a8-365409103274" (UID: "80c21e8c-3128-48ef-91a8-365409103274"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.992366 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.992598 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.992670 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.992795 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80c21e8c-3128-48ef-91a8-365409103274-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:27 crc kubenswrapper[4809]: I1127 17:58:27.992865 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4h22\" (UniqueName: \"kubernetes.io/projected/80c21e8c-3128-48ef-91a8-365409103274-kube-api-access-k4h22\") on node \"crc\" DevicePath \"\"" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.381503 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" event={"ID":"80c21e8c-3128-48ef-91a8-365409103274","Type":"ContainerDied","Data":"d20504ab05d8a60dba3d0f471992a9c9c4f2be76c68f6908ee474bb12135dd5e"} Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.381545 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d20504ab05d8a60dba3d0f471992a9c9c4f2be76c68f6908ee474bb12135dd5e" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.381596 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.468820 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795"] Nov 27 17:58:28 crc kubenswrapper[4809]: E1127 17:58:28.469210 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerName="registry-server" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469223 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerName="registry-server" Nov 27 17:58:28 crc kubenswrapper[4809]: E1127 17:58:28.469255 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerName="registry-server" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469261 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerName="registry-server" Nov 27 17:58:28 crc kubenswrapper[4809]: E1127 17:58:28.469270 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c21e8c-3128-48ef-91a8-365409103274" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469277 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c21e8c-3128-48ef-91a8-365409103274" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 27 17:58:28 crc kubenswrapper[4809]: E1127 17:58:28.469286 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerName="extract-utilities" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469291 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerName="extract-utilities" Nov 27 17:58:28 crc kubenswrapper[4809]: E1127 17:58:28.469311 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerName="extract-utilities" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469316 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerName="extract-utilities" Nov 27 17:58:28 crc kubenswrapper[4809]: E1127 17:58:28.469326 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerName="extract-content" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469331 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerName="extract-content" Nov 27 17:58:28 crc kubenswrapper[4809]: E1127 17:58:28.469346 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerName="extract-content" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469351 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerName="extract-content" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469529 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a55fb83c-f00e-406d-ba12-9e0dd182f98d" containerName="registry-server" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469547 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="80c21e8c-3128-48ef-91a8-365409103274" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.469556 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae66b3c-224c-47e6-88d2-21fc1e9dd437" containerName="registry-server" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.470236 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.473774 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.474033 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.474240 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.474413 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.474569 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.474774 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.475579 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.490661 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795"] Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.604852 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbhsc\" (UniqueName: \"kubernetes.io/projected/9230a20e-bbed-4c09-8e94-a87279b78723-kube-api-access-fbhsc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.605174 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9230a20e-bbed-4c09-8e94-a87279b78723-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.605201 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.605233 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.605453 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.605813 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.605971 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.606077 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.606241 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.707906 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9230a20e-bbed-4c09-8e94-a87279b78723-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.707950 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.707974 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.708002 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.708053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.708079 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.708175 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.708218 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.708946 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9230a20e-bbed-4c09-8e94-a87279b78723-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.709915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbhsc\" (UniqueName: \"kubernetes.io/projected/9230a20e-bbed-4c09-8e94-a87279b78723-kube-api-access-fbhsc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.713465 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.714137 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.714143 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.714208 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.714882 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.715835 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.716312 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.726581 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbhsc\" (UniqueName: \"kubernetes.io/projected/9230a20e-bbed-4c09-8e94-a87279b78723-kube-api-access-fbhsc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-zc795\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:28 crc kubenswrapper[4809]: I1127 17:58:28.788651 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 17:58:29 crc kubenswrapper[4809]: I1127 17:58:29.312557 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795"] Nov 27 17:58:29 crc kubenswrapper[4809]: W1127 17:58:29.316069 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9230a20e_bbed_4c09_8e94_a87279b78723.slice/crio-3529de8137a78b2f6de986e22c410d2bf575ca172b83178ee60280a5522d6d28 WatchSource:0}: Error finding container 3529de8137a78b2f6de986e22c410d2bf575ca172b83178ee60280a5522d6d28: Status 404 returned error can't find the container with id 3529de8137a78b2f6de986e22c410d2bf575ca172b83178ee60280a5522d6d28 Nov 27 17:58:29 crc kubenswrapper[4809]: I1127 17:58:29.393643 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" event={"ID":"9230a20e-bbed-4c09-8e94-a87279b78723","Type":"ContainerStarted","Data":"3529de8137a78b2f6de986e22c410d2bf575ca172b83178ee60280a5522d6d28"} Nov 27 17:58:31 crc kubenswrapper[4809]: I1127 17:58:31.412593 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" event={"ID":"9230a20e-bbed-4c09-8e94-a87279b78723","Type":"ContainerStarted","Data":"eadcdb4080fbf6c13df40bbf276a4a1bd082911d28fff724fd1038a267c9838f"} Nov 27 17:58:31 crc kubenswrapper[4809]: I1127 17:58:31.430573 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" podStartSLOduration=1.829224781 podStartE2EDuration="3.43055276s" podCreationTimestamp="2025-11-27 17:58:28 +0000 UTC" firstStartedPulling="2025-11-27 17:58:29.318689037 +0000 UTC m=+2944.591146389" lastFinishedPulling="2025-11-27 17:58:30.920017016 +0000 UTC m=+2946.192474368" observedRunningTime="2025-11-27 17:58:31.427717592 +0000 UTC m=+2946.700174944" watchObservedRunningTime="2025-11-27 17:58:31.43055276 +0000 UTC m=+2946.703010112" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.147820 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5"] Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.150035 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.152915 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.152982 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.161670 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5"] Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.253625 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae8d97af-2e83-4808-a6cc-765c65b7a19a-config-volume\") pod \"collect-profiles-29404440-vnwk5\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.253715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae8d97af-2e83-4808-a6cc-765c65b7a19a-secret-volume\") pod \"collect-profiles-29404440-vnwk5\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.253785 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm6sj\" (UniqueName: \"kubernetes.io/projected/ae8d97af-2e83-4808-a6cc-765c65b7a19a-kube-api-access-mm6sj\") pod \"collect-profiles-29404440-vnwk5\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.355789 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae8d97af-2e83-4808-a6cc-765c65b7a19a-secret-volume\") pod \"collect-profiles-29404440-vnwk5\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.355901 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm6sj\" (UniqueName: \"kubernetes.io/projected/ae8d97af-2e83-4808-a6cc-765c65b7a19a-kube-api-access-mm6sj\") pod \"collect-profiles-29404440-vnwk5\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.356154 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae8d97af-2e83-4808-a6cc-765c65b7a19a-config-volume\") pod \"collect-profiles-29404440-vnwk5\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.357522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae8d97af-2e83-4808-a6cc-765c65b7a19a-config-volume\") pod \"collect-profiles-29404440-vnwk5\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.365323 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae8d97af-2e83-4808-a6cc-765c65b7a19a-secret-volume\") pod \"collect-profiles-29404440-vnwk5\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.375864 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm6sj\" (UniqueName: \"kubernetes.io/projected/ae8d97af-2e83-4808-a6cc-765c65b7a19a-kube-api-access-mm6sj\") pod \"collect-profiles-29404440-vnwk5\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.480495 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:00 crc kubenswrapper[4809]: I1127 18:00:00.921296 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5"] Nov 27 18:00:01 crc kubenswrapper[4809]: I1127 18:00:01.685718 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae8d97af-2e83-4808-a6cc-765c65b7a19a" containerID="99c4a0724257f6072bcb0f0808a20b7a13b349d1f9e8ffac55e9b03bd68676cb" exitCode=0 Nov 27 18:00:01 crc kubenswrapper[4809]: I1127 18:00:01.685806 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" event={"ID":"ae8d97af-2e83-4808-a6cc-765c65b7a19a","Type":"ContainerDied","Data":"99c4a0724257f6072bcb0f0808a20b7a13b349d1f9e8ffac55e9b03bd68676cb"} Nov 27 18:00:01 crc kubenswrapper[4809]: I1127 18:00:01.686069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" event={"ID":"ae8d97af-2e83-4808-a6cc-765c65b7a19a","Type":"ContainerStarted","Data":"ade644946fa5b5851c271247194a9c34cda75522b3b3b0e14208625dcecf112b"} Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.057017 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.116956 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm6sj\" (UniqueName: \"kubernetes.io/projected/ae8d97af-2e83-4808-a6cc-765c65b7a19a-kube-api-access-mm6sj\") pod \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.117006 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae8d97af-2e83-4808-a6cc-765c65b7a19a-secret-volume\") pod \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.117119 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae8d97af-2e83-4808-a6cc-765c65b7a19a-config-volume\") pod \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\" (UID: \"ae8d97af-2e83-4808-a6cc-765c65b7a19a\") " Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.117920 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae8d97af-2e83-4808-a6cc-765c65b7a19a-config-volume" (OuterVolumeSpecName: "config-volume") pod "ae8d97af-2e83-4808-a6cc-765c65b7a19a" (UID: "ae8d97af-2e83-4808-a6cc-765c65b7a19a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.123963 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8d97af-2e83-4808-a6cc-765c65b7a19a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ae8d97af-2e83-4808-a6cc-765c65b7a19a" (UID: "ae8d97af-2e83-4808-a6cc-765c65b7a19a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.124188 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8d97af-2e83-4808-a6cc-765c65b7a19a-kube-api-access-mm6sj" (OuterVolumeSpecName: "kube-api-access-mm6sj") pod "ae8d97af-2e83-4808-a6cc-765c65b7a19a" (UID: "ae8d97af-2e83-4808-a6cc-765c65b7a19a"). InnerVolumeSpecName "kube-api-access-mm6sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.219888 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm6sj\" (UniqueName: \"kubernetes.io/projected/ae8d97af-2e83-4808-a6cc-765c65b7a19a-kube-api-access-mm6sj\") on node \"crc\" DevicePath \"\"" Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.219925 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae8d97af-2e83-4808-a6cc-765c65b7a19a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.219933 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae8d97af-2e83-4808-a6cc-765c65b7a19a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.704205 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" event={"ID":"ae8d97af-2e83-4808-a6cc-765c65b7a19a","Type":"ContainerDied","Data":"ade644946fa5b5851c271247194a9c34cda75522b3b3b0e14208625dcecf112b"} Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.704256 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ade644946fa5b5851c271247194a9c34cda75522b3b3b0e14208625dcecf112b" Nov 27 18:00:03 crc kubenswrapper[4809]: I1127 18:00:03.704339 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404440-vnwk5" Nov 27 18:00:04 crc kubenswrapper[4809]: I1127 18:00:04.127268 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j"] Nov 27 18:00:04 crc kubenswrapper[4809]: I1127 18:00:04.137313 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404395-d698j"] Nov 27 18:00:05 crc kubenswrapper[4809]: I1127 18:00:05.472059 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3f4af94-6c9c-498a-b1c9-828c7bce2226" path="/var/lib/kubelet/pods/b3f4af94-6c9c-498a-b1c9-828c7bce2226/volumes" Nov 27 18:00:25 crc kubenswrapper[4809]: I1127 18:00:25.779242 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:00:25 crc kubenswrapper[4809]: I1127 18:00:25.779835 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:00:35 crc kubenswrapper[4809]: I1127 18:00:35.101139 4809 scope.go:117] "RemoveContainer" containerID="05a94a0b1200c4080367a07691589fd5a78843b330a83f0c054465ad4ec2f5c1" Nov 27 18:00:55 crc kubenswrapper[4809]: I1127 18:00:55.779067 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:00:55 crc kubenswrapper[4809]: I1127 18:00:55.779504 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.150371 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29404441-hv6j5"] Nov 27 18:01:00 crc kubenswrapper[4809]: E1127 18:01:00.151389 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8d97af-2e83-4808-a6cc-765c65b7a19a" containerName="collect-profiles" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.151402 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8d97af-2e83-4808-a6cc-765c65b7a19a" containerName="collect-profiles" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.151580 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8d97af-2e83-4808-a6cc-765c65b7a19a" containerName="collect-profiles" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.152398 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.161804 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29404441-hv6j5"] Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.327032 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-config-data\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.327390 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-fernet-keys\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.327550 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-combined-ca-bundle\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.327667 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7clb\" (UniqueName: \"kubernetes.io/projected/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-kube-api-access-n7clb\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.429271 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-combined-ca-bundle\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.429347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7clb\" (UniqueName: \"kubernetes.io/projected/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-kube-api-access-n7clb\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.429380 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-config-data\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.429502 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-fernet-keys\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.436431 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-fernet-keys\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.436544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-config-data\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.437511 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-combined-ca-bundle\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.450044 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7clb\" (UniqueName: \"kubernetes.io/projected/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-kube-api-access-n7clb\") pod \"keystone-cron-29404441-hv6j5\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.496045 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:00 crc kubenswrapper[4809]: I1127 18:01:00.932421 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29404441-hv6j5"] Nov 27 18:01:01 crc kubenswrapper[4809]: I1127 18:01:01.229515 4809 generic.go:334] "Generic (PLEG): container finished" podID="9230a20e-bbed-4c09-8e94-a87279b78723" containerID="eadcdb4080fbf6c13df40bbf276a4a1bd082911d28fff724fd1038a267c9838f" exitCode=0 Nov 27 18:01:01 crc kubenswrapper[4809]: I1127 18:01:01.229589 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" event={"ID":"9230a20e-bbed-4c09-8e94-a87279b78723","Type":"ContainerDied","Data":"eadcdb4080fbf6c13df40bbf276a4a1bd082911d28fff724fd1038a267c9838f"} Nov 27 18:01:01 crc kubenswrapper[4809]: I1127 18:01:01.230880 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404441-hv6j5" event={"ID":"47d4ccb6-b0c6-4676-a5d0-eb82e862457d","Type":"ContainerStarted","Data":"9042b46c38ce288a0be2c5decb39d31c83d6524ed452b248cb888b7d83adfa8a"} Nov 27 18:01:01 crc kubenswrapper[4809]: I1127 18:01:01.230907 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404441-hv6j5" event={"ID":"47d4ccb6-b0c6-4676-a5d0-eb82e862457d","Type":"ContainerStarted","Data":"897117e62ff65978612707bea6a72e7fbe6f8aee4e4377d531144d26d87d1bd1"} Nov 27 18:01:01 crc kubenswrapper[4809]: I1127 18:01:01.274820 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29404441-hv6j5" podStartSLOduration=1.274797181 podStartE2EDuration="1.274797181s" podCreationTimestamp="2025-11-27 18:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 18:01:01.270484343 +0000 UTC m=+3096.542941715" watchObservedRunningTime="2025-11-27 18:01:01.274797181 +0000 UTC m=+3096.547254533" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.717628 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.879219 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-ssh-key\") pod \"9230a20e-bbed-4c09-8e94-a87279b78723\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.879333 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-0\") pod \"9230a20e-bbed-4c09-8e94-a87279b78723\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.879409 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbhsc\" (UniqueName: \"kubernetes.io/projected/9230a20e-bbed-4c09-8e94-a87279b78723-kube-api-access-fbhsc\") pod \"9230a20e-bbed-4c09-8e94-a87279b78723\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.879432 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-combined-ca-bundle\") pod \"9230a20e-bbed-4c09-8e94-a87279b78723\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.879472 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-inventory\") pod \"9230a20e-bbed-4c09-8e94-a87279b78723\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.879580 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9230a20e-bbed-4c09-8e94-a87279b78723-nova-extra-config-0\") pod \"9230a20e-bbed-4c09-8e94-a87279b78723\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.879628 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-1\") pod \"9230a20e-bbed-4c09-8e94-a87279b78723\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.879684 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-0\") pod \"9230a20e-bbed-4c09-8e94-a87279b78723\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.880288 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-1\") pod \"9230a20e-bbed-4c09-8e94-a87279b78723\" (UID: \"9230a20e-bbed-4c09-8e94-a87279b78723\") " Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.886791 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9230a20e-bbed-4c09-8e94-a87279b78723-kube-api-access-fbhsc" (OuterVolumeSpecName: "kube-api-access-fbhsc") pod "9230a20e-bbed-4c09-8e94-a87279b78723" (UID: "9230a20e-bbed-4c09-8e94-a87279b78723"). InnerVolumeSpecName "kube-api-access-fbhsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.886966 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "9230a20e-bbed-4c09-8e94-a87279b78723" (UID: "9230a20e-bbed-4c09-8e94-a87279b78723"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.913847 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "9230a20e-bbed-4c09-8e94-a87279b78723" (UID: "9230a20e-bbed-4c09-8e94-a87279b78723"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.914879 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-inventory" (OuterVolumeSpecName: "inventory") pod "9230a20e-bbed-4c09-8e94-a87279b78723" (UID: "9230a20e-bbed-4c09-8e94-a87279b78723"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.919761 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "9230a20e-bbed-4c09-8e94-a87279b78723" (UID: "9230a20e-bbed-4c09-8e94-a87279b78723"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.923318 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9230a20e-bbed-4c09-8e94-a87279b78723-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "9230a20e-bbed-4c09-8e94-a87279b78723" (UID: "9230a20e-bbed-4c09-8e94-a87279b78723"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.924545 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "9230a20e-bbed-4c09-8e94-a87279b78723" (UID: "9230a20e-bbed-4c09-8e94-a87279b78723"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.925061 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "9230a20e-bbed-4c09-8e94-a87279b78723" (UID: "9230a20e-bbed-4c09-8e94-a87279b78723"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.925498 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9230a20e-bbed-4c09-8e94-a87279b78723" (UID: "9230a20e-bbed-4c09-8e94-a87279b78723"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.982805 4809 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9230a20e-bbed-4c09-8e94-a87279b78723-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.982845 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.982855 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.982866 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.982874 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.982883 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.982892 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbhsc\" (UniqueName: \"kubernetes.io/projected/9230a20e-bbed-4c09-8e94-a87279b78723-kube-api-access-fbhsc\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.982900 4809 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:02 crc kubenswrapper[4809]: I1127 18:01:02.982911 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9230a20e-bbed-4c09-8e94-a87279b78723-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.251796 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" event={"ID":"9230a20e-bbed-4c09-8e94-a87279b78723","Type":"ContainerDied","Data":"3529de8137a78b2f6de986e22c410d2bf575ca172b83178ee60280a5522d6d28"} Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.252272 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3529de8137a78b2f6de986e22c410d2bf575ca172b83178ee60280a5522d6d28" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.251891 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-zc795" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.329191 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495"] Nov 27 18:01:03 crc kubenswrapper[4809]: E1127 18:01:03.329896 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9230a20e-bbed-4c09-8e94-a87279b78723" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.329987 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9230a20e-bbed-4c09-8e94-a87279b78723" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.330305 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9230a20e-bbed-4c09-8e94-a87279b78723" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.331131 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.333968 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.334095 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.334384 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.335540 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.335713 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.343515 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495"] Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.495882 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.496072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.496153 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.496254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2r7c\" (UniqueName: \"kubernetes.io/projected/10c284aa-b6df-41bb-8ace-7dd18b804925-kube-api-access-h2r7c\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.496313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.496343 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.496379 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.598484 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2r7c\" (UniqueName: \"kubernetes.io/projected/10c284aa-b6df-41bb-8ace-7dd18b804925-kube-api-access-h2r7c\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.598597 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.598620 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.598651 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.598823 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.598848 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.598904 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.604004 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.604060 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.604390 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.604553 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.605383 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.606530 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.616485 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2r7c\" (UniqueName: \"kubernetes.io/projected/10c284aa-b6df-41bb-8ace-7dd18b804925-kube-api-access-h2r7c\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kv495\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:03 crc kubenswrapper[4809]: I1127 18:01:03.657957 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:01:04 crc kubenswrapper[4809]: I1127 18:01:04.167393 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495"] Nov 27 18:01:04 crc kubenswrapper[4809]: W1127 18:01:04.178158 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10c284aa_b6df_41bb_8ace_7dd18b804925.slice/crio-0967a6bf016459fc604653f8e75204d5c76739e18a3a278c69c76b74aadd25d6 WatchSource:0}: Error finding container 0967a6bf016459fc604653f8e75204d5c76739e18a3a278c69c76b74aadd25d6: Status 404 returned error can't find the container with id 0967a6bf016459fc604653f8e75204d5c76739e18a3a278c69c76b74aadd25d6 Nov 27 18:01:04 crc kubenswrapper[4809]: I1127 18:01:04.182535 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 18:01:04 crc kubenswrapper[4809]: I1127 18:01:04.263787 4809 generic.go:334] "Generic (PLEG): container finished" podID="47d4ccb6-b0c6-4676-a5d0-eb82e862457d" containerID="9042b46c38ce288a0be2c5decb39d31c83d6524ed452b248cb888b7d83adfa8a" exitCode=0 Nov 27 18:01:04 crc kubenswrapper[4809]: I1127 18:01:04.263885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404441-hv6j5" event={"ID":"47d4ccb6-b0c6-4676-a5d0-eb82e862457d","Type":"ContainerDied","Data":"9042b46c38ce288a0be2c5decb39d31c83d6524ed452b248cb888b7d83adfa8a"} Nov 27 18:01:04 crc kubenswrapper[4809]: I1127 18:01:04.265994 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" event={"ID":"10c284aa-b6df-41bb-8ace-7dd18b804925","Type":"ContainerStarted","Data":"0967a6bf016459fc604653f8e75204d5c76739e18a3a278c69c76b74aadd25d6"} Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.652048 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.759205 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-combined-ca-bundle\") pod \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.759291 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7clb\" (UniqueName: \"kubernetes.io/projected/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-kube-api-access-n7clb\") pod \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.759436 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-fernet-keys\") pod \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.759530 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-config-data\") pod \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\" (UID: \"47d4ccb6-b0c6-4676-a5d0-eb82e862457d\") " Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.764894 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-kube-api-access-n7clb" (OuterVolumeSpecName: "kube-api-access-n7clb") pod "47d4ccb6-b0c6-4676-a5d0-eb82e862457d" (UID: "47d4ccb6-b0c6-4676-a5d0-eb82e862457d"). InnerVolumeSpecName "kube-api-access-n7clb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.764997 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "47d4ccb6-b0c6-4676-a5d0-eb82e862457d" (UID: "47d4ccb6-b0c6-4676-a5d0-eb82e862457d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.797756 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47d4ccb6-b0c6-4676-a5d0-eb82e862457d" (UID: "47d4ccb6-b0c6-4676-a5d0-eb82e862457d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.810288 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-config-data" (OuterVolumeSpecName: "config-data") pod "47d4ccb6-b0c6-4676-a5d0-eb82e862457d" (UID: "47d4ccb6-b0c6-4676-a5d0-eb82e862457d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.861711 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7clb\" (UniqueName: \"kubernetes.io/projected/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-kube-api-access-n7clb\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.861787 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.861798 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:05 crc kubenswrapper[4809]: I1127 18:01:05.861806 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d4ccb6-b0c6-4676-a5d0-eb82e862457d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 18:01:06 crc kubenswrapper[4809]: I1127 18:01:06.284223 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" event={"ID":"10c284aa-b6df-41bb-8ace-7dd18b804925","Type":"ContainerStarted","Data":"d22bd54587133b04ef638885ce92c7cf084253c5a80b3357d9ab341e3c770dc4"} Nov 27 18:01:06 crc kubenswrapper[4809]: I1127 18:01:06.286562 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404441-hv6j5" event={"ID":"47d4ccb6-b0c6-4676-a5d0-eb82e862457d","Type":"ContainerDied","Data":"897117e62ff65978612707bea6a72e7fbe6f8aee4e4377d531144d26d87d1bd1"} Nov 27 18:01:06 crc kubenswrapper[4809]: I1127 18:01:06.286608 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404441-hv6j5" Nov 27 18:01:06 crc kubenswrapper[4809]: I1127 18:01:06.286628 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="897117e62ff65978612707bea6a72e7fbe6f8aee4e4377d531144d26d87d1bd1" Nov 27 18:01:06 crc kubenswrapper[4809]: I1127 18:01:06.321913 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" podStartSLOduration=2.401534415 podStartE2EDuration="3.321893379s" podCreationTimestamp="2025-11-27 18:01:03 +0000 UTC" firstStartedPulling="2025-11-27 18:01:04.182307691 +0000 UTC m=+3099.454765043" lastFinishedPulling="2025-11-27 18:01:05.102666655 +0000 UTC m=+3100.375124007" observedRunningTime="2025-11-27 18:01:06.306780957 +0000 UTC m=+3101.579238309" watchObservedRunningTime="2025-11-27 18:01:06.321893379 +0000 UTC m=+3101.594350731" Nov 27 18:01:25 crc kubenswrapper[4809]: I1127 18:01:25.779579 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:01:25 crc kubenswrapper[4809]: I1127 18:01:25.780124 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:01:25 crc kubenswrapper[4809]: I1127 18:01:25.780166 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 18:01:25 crc kubenswrapper[4809]: I1127 18:01:25.780926 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 18:01:25 crc kubenswrapper[4809]: I1127 18:01:25.780978 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" gracePeriod=600 Nov 27 18:01:25 crc kubenswrapper[4809]: E1127 18:01:25.903368 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:01:26 crc kubenswrapper[4809]: E1127 18:01:26.001429 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b63ded2_aa4f_4aba_b3cc_a3965a01036c.slice/crio-2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b63ded2_aa4f_4aba_b3cc_a3965a01036c.slice/crio-conmon-2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064.scope\": RecentStats: unable to find data in memory cache]" Nov 27 18:01:26 crc kubenswrapper[4809]: I1127 18:01:26.660518 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" exitCode=0 Nov 27 18:01:26 crc kubenswrapper[4809]: I1127 18:01:26.660595 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064"} Nov 27 18:01:26 crc kubenswrapper[4809]: I1127 18:01:26.660889 4809 scope.go:117] "RemoveContainer" containerID="b076d299448271c4dfcc1047bb2cf65bc76068eb6330d466a9636e70144ea920" Nov 27 18:01:26 crc kubenswrapper[4809]: I1127 18:01:26.661957 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:01:26 crc kubenswrapper[4809]: E1127 18:01:26.662331 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:01:38 crc kubenswrapper[4809]: I1127 18:01:38.457941 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:01:38 crc kubenswrapper[4809]: E1127 18:01:38.458769 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:01:50 crc kubenswrapper[4809]: I1127 18:01:50.458386 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:01:50 crc kubenswrapper[4809]: E1127 18:01:50.459459 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:02:01 crc kubenswrapper[4809]: I1127 18:02:01.458024 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:02:01 crc kubenswrapper[4809]: E1127 18:02:01.458771 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:02:15 crc kubenswrapper[4809]: I1127 18:02:15.464507 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:02:15 crc kubenswrapper[4809]: E1127 18:02:15.465338 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:02:28 crc kubenswrapper[4809]: I1127 18:02:28.457944 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:02:28 crc kubenswrapper[4809]: E1127 18:02:28.458677 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:02:42 crc kubenswrapper[4809]: I1127 18:02:42.458458 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:02:42 crc kubenswrapper[4809]: E1127 18:02:42.459334 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:02:56 crc kubenswrapper[4809]: I1127 18:02:56.458956 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:02:56 crc kubenswrapper[4809]: E1127 18:02:56.460284 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:03:05 crc kubenswrapper[4809]: I1127 18:03:05.802428 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d4f9n"] Nov 27 18:03:05 crc kubenswrapper[4809]: E1127 18:03:05.803428 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d4ccb6-b0c6-4676-a5d0-eb82e862457d" containerName="keystone-cron" Nov 27 18:03:05 crc kubenswrapper[4809]: I1127 18:03:05.803442 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d4ccb6-b0c6-4676-a5d0-eb82e862457d" containerName="keystone-cron" Nov 27 18:03:05 crc kubenswrapper[4809]: I1127 18:03:05.803657 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d4ccb6-b0c6-4676-a5d0-eb82e862457d" containerName="keystone-cron" Nov 27 18:03:05 crc kubenswrapper[4809]: I1127 18:03:05.805592 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:05 crc kubenswrapper[4809]: I1127 18:03:05.821359 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d4f9n"] Nov 27 18:03:05 crc kubenswrapper[4809]: I1127 18:03:05.989712 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk9gb\" (UniqueName: \"kubernetes.io/projected/21562ef7-7977-4197-a0ff-c3ffc47e26e8-kube-api-access-wk9gb\") pod \"community-operators-d4f9n\" (UID: \"21562ef7-7977-4197-a0ff-c3ffc47e26e8\") " pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:05 crc kubenswrapper[4809]: I1127 18:03:05.989789 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21562ef7-7977-4197-a0ff-c3ffc47e26e8-utilities\") pod \"community-operators-d4f9n\" (UID: \"21562ef7-7977-4197-a0ff-c3ffc47e26e8\") " pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:05 crc kubenswrapper[4809]: I1127 18:03:05.989821 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21562ef7-7977-4197-a0ff-c3ffc47e26e8-catalog-content\") pod \"community-operators-d4f9n\" (UID: \"21562ef7-7977-4197-a0ff-c3ffc47e26e8\") " pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:06 crc kubenswrapper[4809]: I1127 18:03:06.092230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk9gb\" (UniqueName: \"kubernetes.io/projected/21562ef7-7977-4197-a0ff-c3ffc47e26e8-kube-api-access-wk9gb\") pod \"community-operators-d4f9n\" (UID: \"21562ef7-7977-4197-a0ff-c3ffc47e26e8\") " pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:06 crc kubenswrapper[4809]: I1127 18:03:06.092303 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21562ef7-7977-4197-a0ff-c3ffc47e26e8-utilities\") pod \"community-operators-d4f9n\" (UID: \"21562ef7-7977-4197-a0ff-c3ffc47e26e8\") " pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:06 crc kubenswrapper[4809]: I1127 18:03:06.092350 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21562ef7-7977-4197-a0ff-c3ffc47e26e8-catalog-content\") pod \"community-operators-d4f9n\" (UID: \"21562ef7-7977-4197-a0ff-c3ffc47e26e8\") " pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:06 crc kubenswrapper[4809]: I1127 18:03:06.092847 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21562ef7-7977-4197-a0ff-c3ffc47e26e8-utilities\") pod \"community-operators-d4f9n\" (UID: \"21562ef7-7977-4197-a0ff-c3ffc47e26e8\") " pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:06 crc kubenswrapper[4809]: I1127 18:03:06.092928 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21562ef7-7977-4197-a0ff-c3ffc47e26e8-catalog-content\") pod \"community-operators-d4f9n\" (UID: \"21562ef7-7977-4197-a0ff-c3ffc47e26e8\") " pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:06 crc kubenswrapper[4809]: I1127 18:03:06.114381 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk9gb\" (UniqueName: \"kubernetes.io/projected/21562ef7-7977-4197-a0ff-c3ffc47e26e8-kube-api-access-wk9gb\") pod \"community-operators-d4f9n\" (UID: \"21562ef7-7977-4197-a0ff-c3ffc47e26e8\") " pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:06 crc kubenswrapper[4809]: I1127 18:03:06.131298 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:06 crc kubenswrapper[4809]: I1127 18:03:06.619107 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d4f9n"] Nov 27 18:03:07 crc kubenswrapper[4809]: I1127 18:03:07.457979 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:03:07 crc kubenswrapper[4809]: E1127 18:03:07.459008 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:03:07 crc kubenswrapper[4809]: I1127 18:03:07.617494 4809 generic.go:334] "Generic (PLEG): container finished" podID="21562ef7-7977-4197-a0ff-c3ffc47e26e8" containerID="5ed3fa2115d22a1d5eec1382da68b827d583b8bacd3ef7559fccd1d4c30ca75b" exitCode=0 Nov 27 18:03:07 crc kubenswrapper[4809]: I1127 18:03:07.617535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4f9n" event={"ID":"21562ef7-7977-4197-a0ff-c3ffc47e26e8","Type":"ContainerDied","Data":"5ed3fa2115d22a1d5eec1382da68b827d583b8bacd3ef7559fccd1d4c30ca75b"} Nov 27 18:03:07 crc kubenswrapper[4809]: I1127 18:03:07.617558 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4f9n" event={"ID":"21562ef7-7977-4197-a0ff-c3ffc47e26e8","Type":"ContainerStarted","Data":"7a03162d89ad56f167378be8588ce051e67f1d681b51013ab0994d3a758abee2"} Nov 27 18:03:11 crc kubenswrapper[4809]: I1127 18:03:11.657211 4809 generic.go:334] "Generic (PLEG): container finished" podID="21562ef7-7977-4197-a0ff-c3ffc47e26e8" containerID="cbfd7fa522b517b6b34548f4479c35ca4c745610477f7c778cedc46c7e229572" exitCode=0 Nov 27 18:03:11 crc kubenswrapper[4809]: I1127 18:03:11.657755 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4f9n" event={"ID":"21562ef7-7977-4197-a0ff-c3ffc47e26e8","Type":"ContainerDied","Data":"cbfd7fa522b517b6b34548f4479c35ca4c745610477f7c778cedc46c7e229572"} Nov 27 18:03:12 crc kubenswrapper[4809]: I1127 18:03:12.669133 4809 generic.go:334] "Generic (PLEG): container finished" podID="10c284aa-b6df-41bb-8ace-7dd18b804925" containerID="d22bd54587133b04ef638885ce92c7cf084253c5a80b3357d9ab341e3c770dc4" exitCode=0 Nov 27 18:03:12 crc kubenswrapper[4809]: I1127 18:03:12.669229 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" event={"ID":"10c284aa-b6df-41bb-8ace-7dd18b804925","Type":"ContainerDied","Data":"d22bd54587133b04ef638885ce92c7cf084253c5a80b3357d9ab341e3c770dc4"} Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.077336 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.268776 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ssh-key\") pod \"10c284aa-b6df-41bb-8ace-7dd18b804925\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.269134 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2r7c\" (UniqueName: \"kubernetes.io/projected/10c284aa-b6df-41bb-8ace-7dd18b804925-kube-api-access-h2r7c\") pod \"10c284aa-b6df-41bb-8ace-7dd18b804925\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.269170 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-0\") pod \"10c284aa-b6df-41bb-8ace-7dd18b804925\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.269187 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-inventory\") pod \"10c284aa-b6df-41bb-8ace-7dd18b804925\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.269237 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-1\") pod \"10c284aa-b6df-41bb-8ace-7dd18b804925\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.269316 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-telemetry-combined-ca-bundle\") pod \"10c284aa-b6df-41bb-8ace-7dd18b804925\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.269336 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-2\") pod \"10c284aa-b6df-41bb-8ace-7dd18b804925\" (UID: \"10c284aa-b6df-41bb-8ace-7dd18b804925\") " Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.274814 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c284aa-b6df-41bb-8ace-7dd18b804925-kube-api-access-h2r7c" (OuterVolumeSpecName: "kube-api-access-h2r7c") pod "10c284aa-b6df-41bb-8ace-7dd18b804925" (UID: "10c284aa-b6df-41bb-8ace-7dd18b804925"). InnerVolumeSpecName "kube-api-access-h2r7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.274981 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "10c284aa-b6df-41bb-8ace-7dd18b804925" (UID: "10c284aa-b6df-41bb-8ace-7dd18b804925"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.297959 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "10c284aa-b6df-41bb-8ace-7dd18b804925" (UID: "10c284aa-b6df-41bb-8ace-7dd18b804925"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.299252 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-inventory" (OuterVolumeSpecName: "inventory") pod "10c284aa-b6df-41bb-8ace-7dd18b804925" (UID: "10c284aa-b6df-41bb-8ace-7dd18b804925"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.299733 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "10c284aa-b6df-41bb-8ace-7dd18b804925" (UID: "10c284aa-b6df-41bb-8ace-7dd18b804925"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.301700 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "10c284aa-b6df-41bb-8ace-7dd18b804925" (UID: "10c284aa-b6df-41bb-8ace-7dd18b804925"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.308628 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "10c284aa-b6df-41bb-8ace-7dd18b804925" (UID: "10c284aa-b6df-41bb-8ace-7dd18b804925"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.373006 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.373255 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2r7c\" (UniqueName: \"kubernetes.io/projected/10c284aa-b6df-41bb-8ace-7dd18b804925-kube-api-access-h2r7c\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.373345 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.373432 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.373504 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.373577 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.373648 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10c284aa-b6df-41bb-8ace-7dd18b804925-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.690048 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4f9n" event={"ID":"21562ef7-7977-4197-a0ff-c3ffc47e26e8","Type":"ContainerStarted","Data":"33efe28ad6c5bf952645f01247b7b55f32e9588c6fccdbb41abaaf48f5e6e4a2"} Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.691916 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" event={"ID":"10c284aa-b6df-41bb-8ace-7dd18b804925","Type":"ContainerDied","Data":"0967a6bf016459fc604653f8e75204d5c76739e18a3a278c69c76b74aadd25d6"} Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.691970 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0967a6bf016459fc604653f8e75204d5c76739e18a3a278c69c76b74aadd25d6" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.692014 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kv495" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.709858 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d4f9n" podStartSLOduration=3.115603928 podStartE2EDuration="9.709842879s" podCreationTimestamp="2025-11-27 18:03:05 +0000 UTC" firstStartedPulling="2025-11-27 18:03:07.619056345 +0000 UTC m=+3222.891513697" lastFinishedPulling="2025-11-27 18:03:14.213295296 +0000 UTC m=+3229.485752648" observedRunningTime="2025-11-27 18:03:14.708468742 +0000 UTC m=+3229.980926104" watchObservedRunningTime="2025-11-27 18:03:14.709842879 +0000 UTC m=+3229.982300231" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.801298 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s"] Nov 27 18:03:14 crc kubenswrapper[4809]: E1127 18:03:14.801777 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c284aa-b6df-41bb-8ace-7dd18b804925" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.801801 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c284aa-b6df-41bb-8ace-7dd18b804925" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.802074 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c284aa-b6df-41bb-8ace-7dd18b804925" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.802957 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.805883 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.806136 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.806263 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.806393 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.806645 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2fdql" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.812455 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s"] Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.986616 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b2mh\" (UniqueName: \"kubernetes.io/projected/565ebe47-4144-4730-9e10-ebd98010c9a3-kube-api-access-7b2mh\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.986998 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.987123 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.987155 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:14 crc kubenswrapper[4809]: I1127 18:03:14.987184 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.089434 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.089507 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.090385 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b2mh\" (UniqueName: \"kubernetes.io/projected/565ebe47-4144-4730-9e10-ebd98010c9a3-kube-api-access-7b2mh\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.090424 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.090508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.096444 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.096477 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.097064 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.097372 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.110443 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b2mh\" (UniqueName: \"kubernetes.io/projected/565ebe47-4144-4730-9e10-ebd98010c9a3-kube-api-access-7b2mh\") pod \"logging-edpm-deployment-openstack-edpm-ipam-ddw9s\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.119856 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.647421 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s"] Nov 27 18:03:15 crc kubenswrapper[4809]: I1127 18:03:15.705450 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" event={"ID":"565ebe47-4144-4730-9e10-ebd98010c9a3","Type":"ContainerStarted","Data":"f16e4cb52a7fa0f3b21204f6893c3faa2e856b578d2c7c676e4d9e19f44136e6"} Nov 27 18:03:16 crc kubenswrapper[4809]: I1127 18:03:16.132243 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:16 crc kubenswrapper[4809]: I1127 18:03:16.132297 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:16 crc kubenswrapper[4809]: I1127 18:03:16.714858 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" event={"ID":"565ebe47-4144-4730-9e10-ebd98010c9a3","Type":"ContainerStarted","Data":"fbdf48635908475868abc8f27f52f3dabfc8ee94d745d8d6b4533fce143a24c2"} Nov 27 18:03:16 crc kubenswrapper[4809]: I1127 18:03:16.753242 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" podStartSLOduration=2.23890303 podStartE2EDuration="2.753213807s" podCreationTimestamp="2025-11-27 18:03:14 +0000 UTC" firstStartedPulling="2025-11-27 18:03:15.6561708 +0000 UTC m=+3230.928628152" lastFinishedPulling="2025-11-27 18:03:16.170481577 +0000 UTC m=+3231.442938929" observedRunningTime="2025-11-27 18:03:16.739783051 +0000 UTC m=+3232.012240403" watchObservedRunningTime="2025-11-27 18:03:16.753213807 +0000 UTC m=+3232.025671159" Nov 27 18:03:17 crc kubenswrapper[4809]: I1127 18:03:17.176160 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d4f9n" podUID="21562ef7-7977-4197-a0ff-c3ffc47e26e8" containerName="registry-server" probeResult="failure" output=< Nov 27 18:03:17 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 27 18:03:17 crc kubenswrapper[4809]: > Nov 27 18:03:22 crc kubenswrapper[4809]: I1127 18:03:22.457508 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:03:22 crc kubenswrapper[4809]: E1127 18:03:22.458401 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:03:26 crc kubenswrapper[4809]: I1127 18:03:26.193563 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:26 crc kubenswrapper[4809]: I1127 18:03:26.242678 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d4f9n" Nov 27 18:03:26 crc kubenswrapper[4809]: I1127 18:03:26.305672 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d4f9n"] Nov 27 18:03:26 crc kubenswrapper[4809]: I1127 18:03:26.441507 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-28jvg"] Nov 27 18:03:26 crc kubenswrapper[4809]: I1127 18:03:26.441859 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-28jvg" podUID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerName="registry-server" containerID="cri-o://4a4313aa659ffaca6983a372b89dc53a5631a96fe32af46ce252e3d35fc6392c" gracePeriod=2 Nov 27 18:03:26 crc kubenswrapper[4809]: I1127 18:03:26.807074 4809 generic.go:334] "Generic (PLEG): container finished" podID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerID="4a4313aa659ffaca6983a372b89dc53a5631a96fe32af46ce252e3d35fc6392c" exitCode=0 Nov 27 18:03:26 crc kubenswrapper[4809]: I1127 18:03:26.807154 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28jvg" event={"ID":"492f5beb-f94c-4510-a07b-306ef7d369b8","Type":"ContainerDied","Data":"4a4313aa659ffaca6983a372b89dc53a5631a96fe32af46ce252e3d35fc6392c"} Nov 27 18:03:26 crc kubenswrapper[4809]: I1127 18:03:26.909108 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-28jvg" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.026607 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpp5s\" (UniqueName: \"kubernetes.io/projected/492f5beb-f94c-4510-a07b-306ef7d369b8-kube-api-access-zpp5s\") pod \"492f5beb-f94c-4510-a07b-306ef7d369b8\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.026853 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-utilities\") pod \"492f5beb-f94c-4510-a07b-306ef7d369b8\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.026897 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-catalog-content\") pod \"492f5beb-f94c-4510-a07b-306ef7d369b8\" (UID: \"492f5beb-f94c-4510-a07b-306ef7d369b8\") " Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.027863 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-utilities" (OuterVolumeSpecName: "utilities") pod "492f5beb-f94c-4510-a07b-306ef7d369b8" (UID: "492f5beb-f94c-4510-a07b-306ef7d369b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.028224 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.033311 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/492f5beb-f94c-4510-a07b-306ef7d369b8-kube-api-access-zpp5s" (OuterVolumeSpecName: "kube-api-access-zpp5s") pod "492f5beb-f94c-4510-a07b-306ef7d369b8" (UID: "492f5beb-f94c-4510-a07b-306ef7d369b8"). InnerVolumeSpecName "kube-api-access-zpp5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.082080 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "492f5beb-f94c-4510-a07b-306ef7d369b8" (UID: "492f5beb-f94c-4510-a07b-306ef7d369b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.130643 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpp5s\" (UniqueName: \"kubernetes.io/projected/492f5beb-f94c-4510-a07b-306ef7d369b8-kube-api-access-zpp5s\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.130698 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/492f5beb-f94c-4510-a07b-306ef7d369b8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.834658 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-28jvg" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.834823 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28jvg" event={"ID":"492f5beb-f94c-4510-a07b-306ef7d369b8","Type":"ContainerDied","Data":"ea2084e6d4c5e4c504dd08683eee6ee3b3c2686b792df2955a8480e2442e3670"} Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.835986 4809 scope.go:117] "RemoveContainer" containerID="4a4313aa659ffaca6983a372b89dc53a5631a96fe32af46ce252e3d35fc6392c" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.865230 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-28jvg"] Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.871716 4809 scope.go:117] "RemoveContainer" containerID="16b24f37185280255bcacedf08aaec7ce277eccc2b2b0c3e54a89599bc98fbe4" Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.880033 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-28jvg"] Nov 27 18:03:27 crc kubenswrapper[4809]: I1127 18:03:27.901850 4809 scope.go:117] "RemoveContainer" containerID="62db8d1d2f0dea49b01f5d278ddbe593a19f389b050682dd0aea3f4ef9ffeb39" Nov 27 18:03:29 crc kubenswrapper[4809]: I1127 18:03:29.468412 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="492f5beb-f94c-4510-a07b-306ef7d369b8" path="/var/lib/kubelet/pods/492f5beb-f94c-4510-a07b-306ef7d369b8/volumes" Nov 27 18:03:29 crc kubenswrapper[4809]: I1127 18:03:29.854208 4809 generic.go:334] "Generic (PLEG): container finished" podID="565ebe47-4144-4730-9e10-ebd98010c9a3" containerID="fbdf48635908475868abc8f27f52f3dabfc8ee94d745d8d6b4533fce143a24c2" exitCode=0 Nov 27 18:03:29 crc kubenswrapper[4809]: I1127 18:03:29.854250 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" event={"ID":"565ebe47-4144-4730-9e10-ebd98010c9a3","Type":"ContainerDied","Data":"fbdf48635908475868abc8f27f52f3dabfc8ee94d745d8d6b4533fce143a24c2"} Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.256187 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.415771 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-1\") pod \"565ebe47-4144-4730-9e10-ebd98010c9a3\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.416244 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b2mh\" (UniqueName: \"kubernetes.io/projected/565ebe47-4144-4730-9e10-ebd98010c9a3-kube-api-access-7b2mh\") pod \"565ebe47-4144-4730-9e10-ebd98010c9a3\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.416371 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-0\") pod \"565ebe47-4144-4730-9e10-ebd98010c9a3\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.416922 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-ssh-key\") pod \"565ebe47-4144-4730-9e10-ebd98010c9a3\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.417005 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-inventory\") pod \"565ebe47-4144-4730-9e10-ebd98010c9a3\" (UID: \"565ebe47-4144-4730-9e10-ebd98010c9a3\") " Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.423095 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565ebe47-4144-4730-9e10-ebd98010c9a3-kube-api-access-7b2mh" (OuterVolumeSpecName: "kube-api-access-7b2mh") pod "565ebe47-4144-4730-9e10-ebd98010c9a3" (UID: "565ebe47-4144-4730-9e10-ebd98010c9a3"). InnerVolumeSpecName "kube-api-access-7b2mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.448663 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "565ebe47-4144-4730-9e10-ebd98010c9a3" (UID: "565ebe47-4144-4730-9e10-ebd98010c9a3"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.449045 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-inventory" (OuterVolumeSpecName: "inventory") pod "565ebe47-4144-4730-9e10-ebd98010c9a3" (UID: "565ebe47-4144-4730-9e10-ebd98010c9a3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.453502 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "565ebe47-4144-4730-9e10-ebd98010c9a3" (UID: "565ebe47-4144-4730-9e10-ebd98010c9a3"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.459557 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "565ebe47-4144-4730-9e10-ebd98010c9a3" (UID: "565ebe47-4144-4730-9e10-ebd98010c9a3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.519720 4809 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.519783 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.519799 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.519810 4809 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/565ebe47-4144-4730-9e10-ebd98010c9a3-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.519824 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b2mh\" (UniqueName: \"kubernetes.io/projected/565ebe47-4144-4730-9e10-ebd98010c9a3-kube-api-access-7b2mh\") on node \"crc\" DevicePath \"\"" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.871643 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" event={"ID":"565ebe47-4144-4730-9e10-ebd98010c9a3","Type":"ContainerDied","Data":"f16e4cb52a7fa0f3b21204f6893c3faa2e856b578d2c7c676e4d9e19f44136e6"} Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.871676 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-ddw9s" Nov 27 18:03:31 crc kubenswrapper[4809]: I1127 18:03:31.871683 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f16e4cb52a7fa0f3b21204f6893c3faa2e856b578d2c7c676e4d9e19f44136e6" Nov 27 18:03:35 crc kubenswrapper[4809]: I1127 18:03:35.464452 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:03:35 crc kubenswrapper[4809]: E1127 18:03:35.465127 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:03:46 crc kubenswrapper[4809]: I1127 18:03:46.457857 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:03:46 crc kubenswrapper[4809]: E1127 18:03:46.458796 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:03:58 crc kubenswrapper[4809]: I1127 18:03:58.457681 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:03:58 crc kubenswrapper[4809]: E1127 18:03:58.458587 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:04:13 crc kubenswrapper[4809]: I1127 18:04:13.457877 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:04:13 crc kubenswrapper[4809]: E1127 18:04:13.458597 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:04:25 crc kubenswrapper[4809]: I1127 18:04:25.478716 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:04:25 crc kubenswrapper[4809]: E1127 18:04:25.479529 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:04:36 crc kubenswrapper[4809]: I1127 18:04:36.458751 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:04:36 crc kubenswrapper[4809]: E1127 18:04:36.459666 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:04:49 crc kubenswrapper[4809]: I1127 18:04:49.458860 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:04:49 crc kubenswrapper[4809]: E1127 18:04:49.459496 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:05:01 crc kubenswrapper[4809]: I1127 18:05:01.458034 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:05:01 crc kubenswrapper[4809]: E1127 18:05:01.458773 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:05:14 crc kubenswrapper[4809]: I1127 18:05:14.457618 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:05:14 crc kubenswrapper[4809]: E1127 18:05:14.458618 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:05:26 crc kubenswrapper[4809]: I1127 18:05:26.458544 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:05:26 crc kubenswrapper[4809]: E1127 18:05:26.459341 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:05:40 crc kubenswrapper[4809]: I1127 18:05:40.458328 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:05:40 crc kubenswrapper[4809]: E1127 18:05:40.459694 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:05:53 crc kubenswrapper[4809]: I1127 18:05:53.457798 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:05:53 crc kubenswrapper[4809]: E1127 18:05:53.458525 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:06:05 crc kubenswrapper[4809]: I1127 18:06:05.465369 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:06:05 crc kubenswrapper[4809]: E1127 18:06:05.466356 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.443790 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 27 18:06:16 crc kubenswrapper[4809]: E1127 18:06:16.444824 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerName="registry-server" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.444841 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerName="registry-server" Nov 27 18:06:16 crc kubenswrapper[4809]: E1127 18:06:16.444860 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565ebe47-4144-4730-9e10-ebd98010c9a3" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.444867 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="565ebe47-4144-4730-9e10-ebd98010c9a3" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 27 18:06:16 crc kubenswrapper[4809]: E1127 18:06:16.444899 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerName="extract-content" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.444905 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerName="extract-content" Nov 27 18:06:16 crc kubenswrapper[4809]: E1127 18:06:16.444915 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerName="extract-utilities" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.444921 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerName="extract-utilities" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.445138 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="492f5beb-f94c-4510-a07b-306ef7d369b8" containerName="registry-server" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.445157 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="565ebe47-4144-4730-9e10-ebd98010c9a3" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.445847 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.448383 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.448382 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.448627 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4g7qk" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.449342 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.465191 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.602771 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.602822 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.603015 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-config-data\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.603217 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.603331 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.603409 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qxsl\" (UniqueName: \"kubernetes.io/projected/79f6ceba-551a-427c-8690-1c8db833367d-kube-api-access-8qxsl\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.603441 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.603479 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.603996 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.705973 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706063 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qxsl\" (UniqueName: \"kubernetes.io/projected/79f6ceba-551a-427c-8690-1c8db833367d-kube-api-access-8qxsl\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706081 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706100 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706170 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706240 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706256 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706289 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-config-data\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706518 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706792 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.706917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.707588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-config-data\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.707834 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.713071 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.714429 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.714963 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.723950 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qxsl\" (UniqueName: \"kubernetes.io/projected/79f6ceba-551a-427c-8690-1c8db833367d-kube-api-access-8qxsl\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.735768 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " pod="openstack/tempest-tests-tempest" Nov 27 18:06:16 crc kubenswrapper[4809]: I1127 18:06:16.772629 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 27 18:06:17 crc kubenswrapper[4809]: I1127 18:06:17.200074 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 27 18:06:17 crc kubenswrapper[4809]: I1127 18:06:17.204616 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 18:06:17 crc kubenswrapper[4809]: I1127 18:06:17.324650 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"79f6ceba-551a-427c-8690-1c8db833367d","Type":"ContainerStarted","Data":"969d03f886578c0ca962b37aa51fc829faf08801b1a2a0c6b89003f4044a51b3"} Nov 27 18:06:18 crc kubenswrapper[4809]: I1127 18:06:18.463010 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:06:18 crc kubenswrapper[4809]: E1127 18:06:18.463670 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:06:30 crc kubenswrapper[4809]: I1127 18:06:30.458441 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:06:45 crc kubenswrapper[4809]: E1127 18:06:45.034256 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 27 18:06:45 crc kubenswrapper[4809]: E1127 18:06:45.035556 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8qxsl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(79f6ceba-551a-427c-8690-1c8db833367d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 18:06:45 crc kubenswrapper[4809]: E1127 18:06:45.036778 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="79f6ceba-551a-427c-8690-1c8db833367d" Nov 27 18:06:45 crc kubenswrapper[4809]: I1127 18:06:45.654257 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"fb0a07bf48201ad85b3541400956b6a851a810403563fcd255ca613a34ae9a04"} Nov 27 18:06:45 crc kubenswrapper[4809]: E1127 18:06:45.656012 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="79f6ceba-551a-427c-8690-1c8db833367d" Nov 27 18:06:59 crc kubenswrapper[4809]: I1127 18:06:59.529865 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 27 18:07:00 crc kubenswrapper[4809]: I1127 18:07:00.797209 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"79f6ceba-551a-427c-8690-1c8db833367d","Type":"ContainerStarted","Data":"a156fa4fde29499a948a0620b7ef79f077f40346ad257a85ca2465823885e5b3"} Nov 27 18:07:00 crc kubenswrapper[4809]: I1127 18:07:00.837272 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.514719842 podStartE2EDuration="45.83724249s" podCreationTimestamp="2025-11-27 18:06:15 +0000 UTC" firstStartedPulling="2025-11-27 18:06:17.20441511 +0000 UTC m=+3412.476872462" lastFinishedPulling="2025-11-27 18:06:59.526937758 +0000 UTC m=+3454.799395110" observedRunningTime="2025-11-27 18:07:00.822112027 +0000 UTC m=+3456.094569379" watchObservedRunningTime="2025-11-27 18:07:00.83724249 +0000 UTC m=+3456.109699842" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.170766 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6jt79"] Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.178752 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.190557 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6jt79"] Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.382041 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-utilities\") pod \"certified-operators-6jt79\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.382327 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqr7h\" (UniqueName: \"kubernetes.io/projected/ba76bd90-8a17-4830-bd33-f32627b21b91-kube-api-access-tqr7h\") pod \"certified-operators-6jt79\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.382422 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-catalog-content\") pod \"certified-operators-6jt79\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.484363 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqr7h\" (UniqueName: \"kubernetes.io/projected/ba76bd90-8a17-4830-bd33-f32627b21b91-kube-api-access-tqr7h\") pod \"certified-operators-6jt79\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.484515 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-catalog-content\") pod \"certified-operators-6jt79\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.484597 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-utilities\") pod \"certified-operators-6jt79\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.484974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-catalog-content\") pod \"certified-operators-6jt79\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.485033 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-utilities\") pod \"certified-operators-6jt79\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.504961 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqr7h\" (UniqueName: \"kubernetes.io/projected/ba76bd90-8a17-4830-bd33-f32627b21b91-kube-api-access-tqr7h\") pod \"certified-operators-6jt79\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:23 crc kubenswrapper[4809]: I1127 18:07:23.799483 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:24 crc kubenswrapper[4809]: I1127 18:07:24.390577 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6jt79"] Nov 27 18:07:25 crc kubenswrapper[4809]: I1127 18:07:25.039417 4809 generic.go:334] "Generic (PLEG): container finished" podID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerID="675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e" exitCode=0 Nov 27 18:07:25 crc kubenswrapper[4809]: I1127 18:07:25.039488 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jt79" event={"ID":"ba76bd90-8a17-4830-bd33-f32627b21b91","Type":"ContainerDied","Data":"675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e"} Nov 27 18:07:25 crc kubenswrapper[4809]: I1127 18:07:25.040099 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jt79" event={"ID":"ba76bd90-8a17-4830-bd33-f32627b21b91","Type":"ContainerStarted","Data":"e6e70a4c67345a05db4c4d8fc181bad0c538106b49fc6b030429a23a03b736d8"} Nov 27 18:07:27 crc kubenswrapper[4809]: I1127 18:07:27.071553 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jt79" event={"ID":"ba76bd90-8a17-4830-bd33-f32627b21b91","Type":"ContainerStarted","Data":"87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039"} Nov 27 18:07:28 crc kubenswrapper[4809]: I1127 18:07:28.082683 4809 generic.go:334] "Generic (PLEG): container finished" podID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerID="87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039" exitCode=0 Nov 27 18:07:28 crc kubenswrapper[4809]: I1127 18:07:28.082783 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jt79" event={"ID":"ba76bd90-8a17-4830-bd33-f32627b21b91","Type":"ContainerDied","Data":"87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039"} Nov 27 18:07:29 crc kubenswrapper[4809]: I1127 18:07:29.095732 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jt79" event={"ID":"ba76bd90-8a17-4830-bd33-f32627b21b91","Type":"ContainerStarted","Data":"b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2"} Nov 27 18:07:29 crc kubenswrapper[4809]: I1127 18:07:29.122926 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6jt79" podStartSLOduration=2.33896612 podStartE2EDuration="6.122907134s" podCreationTimestamp="2025-11-27 18:07:23 +0000 UTC" firstStartedPulling="2025-11-27 18:07:25.041907962 +0000 UTC m=+3480.314365314" lastFinishedPulling="2025-11-27 18:07:28.825848976 +0000 UTC m=+3484.098306328" observedRunningTime="2025-11-27 18:07:29.116078268 +0000 UTC m=+3484.388535620" watchObservedRunningTime="2025-11-27 18:07:29.122907134 +0000 UTC m=+3484.395364486" Nov 27 18:07:33 crc kubenswrapper[4809]: I1127 18:07:33.799718 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:33 crc kubenswrapper[4809]: I1127 18:07:33.800845 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:33 crc kubenswrapper[4809]: I1127 18:07:33.859326 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:34 crc kubenswrapper[4809]: I1127 18:07:34.197691 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:34 crc kubenswrapper[4809]: I1127 18:07:34.257986 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6jt79"] Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.162040 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6jt79" podUID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerName="registry-server" containerID="cri-o://b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2" gracePeriod=2 Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.653758 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.771474 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqr7h\" (UniqueName: \"kubernetes.io/projected/ba76bd90-8a17-4830-bd33-f32627b21b91-kube-api-access-tqr7h\") pod \"ba76bd90-8a17-4830-bd33-f32627b21b91\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.771696 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-catalog-content\") pod \"ba76bd90-8a17-4830-bd33-f32627b21b91\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.780540 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba76bd90-8a17-4830-bd33-f32627b21b91-kube-api-access-tqr7h" (OuterVolumeSpecName: "kube-api-access-tqr7h") pod "ba76bd90-8a17-4830-bd33-f32627b21b91" (UID: "ba76bd90-8a17-4830-bd33-f32627b21b91"). InnerVolumeSpecName "kube-api-access-tqr7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.781854 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-utilities\") pod \"ba76bd90-8a17-4830-bd33-f32627b21b91\" (UID: \"ba76bd90-8a17-4830-bd33-f32627b21b91\") " Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.782542 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-utilities" (OuterVolumeSpecName: "utilities") pod "ba76bd90-8a17-4830-bd33-f32627b21b91" (UID: "ba76bd90-8a17-4830-bd33-f32627b21b91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.783336 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.783369 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqr7h\" (UniqueName: \"kubernetes.io/projected/ba76bd90-8a17-4830-bd33-f32627b21b91-kube-api-access-tqr7h\") on node \"crc\" DevicePath \"\"" Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.813642 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba76bd90-8a17-4830-bd33-f32627b21b91" (UID: "ba76bd90-8a17-4830-bd33-f32627b21b91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:07:36 crc kubenswrapper[4809]: I1127 18:07:36.885845 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba76bd90-8a17-4830-bd33-f32627b21b91-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.172012 4809 generic.go:334] "Generic (PLEG): container finished" podID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerID="b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2" exitCode=0 Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.172056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jt79" event={"ID":"ba76bd90-8a17-4830-bd33-f32627b21b91","Type":"ContainerDied","Data":"b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2"} Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.172090 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jt79" event={"ID":"ba76bd90-8a17-4830-bd33-f32627b21b91","Type":"ContainerDied","Data":"e6e70a4c67345a05db4c4d8fc181bad0c538106b49fc6b030429a23a03b736d8"} Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.172123 4809 scope.go:117] "RemoveContainer" containerID="b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.172271 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jt79" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.217190 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6jt79"] Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.217485 4809 scope.go:117] "RemoveContainer" containerID="87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.227717 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6jt79"] Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.254140 4809 scope.go:117] "RemoveContainer" containerID="675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.324351 4809 scope.go:117] "RemoveContainer" containerID="b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2" Nov 27 18:07:37 crc kubenswrapper[4809]: E1127 18:07:37.326256 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2\": container with ID starting with b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2 not found: ID does not exist" containerID="b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.326298 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2"} err="failed to get container status \"b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2\": rpc error: code = NotFound desc = could not find container \"b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2\": container with ID starting with b7d518a1b09649b0c39af871c4b286cdff5582f498e704437f75b535885ec9d2 not found: ID does not exist" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.326326 4809 scope.go:117] "RemoveContainer" containerID="87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039" Nov 27 18:07:37 crc kubenswrapper[4809]: E1127 18:07:37.326989 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039\": container with ID starting with 87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039 not found: ID does not exist" containerID="87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.327036 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039"} err="failed to get container status \"87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039\": rpc error: code = NotFound desc = could not find container \"87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039\": container with ID starting with 87c86ed112786990c57849c4d9cb0891c32cd8809962424ca8f350eaa68fe039 not found: ID does not exist" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.327076 4809 scope.go:117] "RemoveContainer" containerID="675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e" Nov 27 18:07:37 crc kubenswrapper[4809]: E1127 18:07:37.327585 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e\": container with ID starting with 675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e not found: ID does not exist" containerID="675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.327610 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e"} err="failed to get container status \"675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e\": rpc error: code = NotFound desc = could not find container \"675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e\": container with ID starting with 675a5da7fde85d7180917c56839159343724207b9d3d855dba50390f3e54176e not found: ID does not exist" Nov 27 18:07:37 crc kubenswrapper[4809]: I1127 18:07:37.469630 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba76bd90-8a17-4830-bd33-f32627b21b91" path="/var/lib/kubelet/pods/ba76bd90-8a17-4830-bd33-f32627b21b91/volumes" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.472117 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kb7ck"] Nov 27 18:08:02 crc kubenswrapper[4809]: E1127 18:08:02.473148 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerName="extract-content" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.473162 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerName="extract-content" Nov 27 18:08:02 crc kubenswrapper[4809]: E1127 18:08:02.473191 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerName="extract-utilities" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.473197 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerName="extract-utilities" Nov 27 18:08:02 crc kubenswrapper[4809]: E1127 18:08:02.473219 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerName="registry-server" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.473226 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerName="registry-server" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.473421 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba76bd90-8a17-4830-bd33-f32627b21b91" containerName="registry-server" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.475118 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.485843 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kb7ck"] Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.663243 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-catalog-content\") pod \"redhat-operators-kb7ck\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.664211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ws5k\" (UniqueName: \"kubernetes.io/projected/8d92e18f-7194-48c3-8db7-7f86b03289ec-kube-api-access-8ws5k\") pod \"redhat-operators-kb7ck\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.664630 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-utilities\") pod \"redhat-operators-kb7ck\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.766951 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-catalog-content\") pod \"redhat-operators-kb7ck\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.767088 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ws5k\" (UniqueName: \"kubernetes.io/projected/8d92e18f-7194-48c3-8db7-7f86b03289ec-kube-api-access-8ws5k\") pod \"redhat-operators-kb7ck\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.767209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-utilities\") pod \"redhat-operators-kb7ck\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.767494 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-catalog-content\") pod \"redhat-operators-kb7ck\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.767568 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-utilities\") pod \"redhat-operators-kb7ck\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.801036 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ws5k\" (UniqueName: \"kubernetes.io/projected/8d92e18f-7194-48c3-8db7-7f86b03289ec-kube-api-access-8ws5k\") pod \"redhat-operators-kb7ck\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:02 crc kubenswrapper[4809]: I1127 18:08:02.801623 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:03 crc kubenswrapper[4809]: I1127 18:08:03.300109 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kb7ck"] Nov 27 18:08:03 crc kubenswrapper[4809]: I1127 18:08:03.433188 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb7ck" event={"ID":"8d92e18f-7194-48c3-8db7-7f86b03289ec","Type":"ContainerStarted","Data":"2ff029cd2a22f3540444e23cf361d008148517c4426389544505183f14141f88"} Nov 27 18:08:04 crc kubenswrapper[4809]: I1127 18:08:04.446616 4809 generic.go:334] "Generic (PLEG): container finished" podID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerID="c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140" exitCode=0 Nov 27 18:08:04 crc kubenswrapper[4809]: I1127 18:08:04.446816 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb7ck" event={"ID":"8d92e18f-7194-48c3-8db7-7f86b03289ec","Type":"ContainerDied","Data":"c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140"} Nov 27 18:08:06 crc kubenswrapper[4809]: I1127 18:08:06.472619 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb7ck" event={"ID":"8d92e18f-7194-48c3-8db7-7f86b03289ec","Type":"ContainerStarted","Data":"a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab"} Nov 27 18:08:09 crc kubenswrapper[4809]: I1127 18:08:09.505252 4809 generic.go:334] "Generic (PLEG): container finished" podID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerID="a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab" exitCode=0 Nov 27 18:08:09 crc kubenswrapper[4809]: I1127 18:08:09.505337 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb7ck" event={"ID":"8d92e18f-7194-48c3-8db7-7f86b03289ec","Type":"ContainerDied","Data":"a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab"} Nov 27 18:08:10 crc kubenswrapper[4809]: I1127 18:08:10.521945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb7ck" event={"ID":"8d92e18f-7194-48c3-8db7-7f86b03289ec","Type":"ContainerStarted","Data":"8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5"} Nov 27 18:08:10 crc kubenswrapper[4809]: I1127 18:08:10.543134 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kb7ck" podStartSLOduration=2.824950762 podStartE2EDuration="8.543117336s" podCreationTimestamp="2025-11-27 18:08:02 +0000 UTC" firstStartedPulling="2025-11-27 18:08:04.449219687 +0000 UTC m=+3519.721677039" lastFinishedPulling="2025-11-27 18:08:10.167386261 +0000 UTC m=+3525.439843613" observedRunningTime="2025-11-27 18:08:10.538980962 +0000 UTC m=+3525.811438314" watchObservedRunningTime="2025-11-27 18:08:10.543117336 +0000 UTC m=+3525.815574688" Nov 27 18:08:12 crc kubenswrapper[4809]: I1127 18:08:12.802192 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:12 crc kubenswrapper[4809]: I1127 18:08:12.802808 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:13 crc kubenswrapper[4809]: I1127 18:08:13.869084 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kb7ck" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerName="registry-server" probeResult="failure" output=< Nov 27 18:08:13 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 27 18:08:13 crc kubenswrapper[4809]: > Nov 27 18:08:22 crc kubenswrapper[4809]: I1127 18:08:22.856087 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:22 crc kubenswrapper[4809]: I1127 18:08:22.911046 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:23 crc kubenswrapper[4809]: I1127 18:08:23.096531 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kb7ck"] Nov 27 18:08:24 crc kubenswrapper[4809]: I1127 18:08:24.656870 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kb7ck" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerName="registry-server" containerID="cri-o://8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5" gracePeriod=2 Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.368894 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.559167 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ws5k\" (UniqueName: \"kubernetes.io/projected/8d92e18f-7194-48c3-8db7-7f86b03289ec-kube-api-access-8ws5k\") pod \"8d92e18f-7194-48c3-8db7-7f86b03289ec\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.559312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-catalog-content\") pod \"8d92e18f-7194-48c3-8db7-7f86b03289ec\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.559414 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-utilities\") pod \"8d92e18f-7194-48c3-8db7-7f86b03289ec\" (UID: \"8d92e18f-7194-48c3-8db7-7f86b03289ec\") " Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.560042 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-utilities" (OuterVolumeSpecName: "utilities") pod "8d92e18f-7194-48c3-8db7-7f86b03289ec" (UID: "8d92e18f-7194-48c3-8db7-7f86b03289ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.560863 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.565018 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d92e18f-7194-48c3-8db7-7f86b03289ec-kube-api-access-8ws5k" (OuterVolumeSpecName: "kube-api-access-8ws5k") pod "8d92e18f-7194-48c3-8db7-7f86b03289ec" (UID: "8d92e18f-7194-48c3-8db7-7f86b03289ec"). InnerVolumeSpecName "kube-api-access-8ws5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.662515 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ws5k\" (UniqueName: \"kubernetes.io/projected/8d92e18f-7194-48c3-8db7-7f86b03289ec-kube-api-access-8ws5k\") on node \"crc\" DevicePath \"\"" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.667463 4809 generic.go:334] "Generic (PLEG): container finished" podID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerID="8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5" exitCode=0 Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.667503 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb7ck" event={"ID":"8d92e18f-7194-48c3-8db7-7f86b03289ec","Type":"ContainerDied","Data":"8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5"} Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.667528 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb7ck" event={"ID":"8d92e18f-7194-48c3-8db7-7f86b03289ec","Type":"ContainerDied","Data":"2ff029cd2a22f3540444e23cf361d008148517c4426389544505183f14141f88"} Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.667543 4809 scope.go:117] "RemoveContainer" containerID="8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.667696 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb7ck" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.682990 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d92e18f-7194-48c3-8db7-7f86b03289ec" (UID: "8d92e18f-7194-48c3-8db7-7f86b03289ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.695619 4809 scope.go:117] "RemoveContainer" containerID="a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.730458 4809 scope.go:117] "RemoveContainer" containerID="c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.765031 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d92e18f-7194-48c3-8db7-7f86b03289ec-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.773381 4809 scope.go:117] "RemoveContainer" containerID="8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5" Nov 27 18:08:25 crc kubenswrapper[4809]: E1127 18:08:25.774379 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5\": container with ID starting with 8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5 not found: ID does not exist" containerID="8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.774422 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5"} err="failed to get container status \"8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5\": rpc error: code = NotFound desc = could not find container \"8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5\": container with ID starting with 8e242d59e48957cfa97b445f3d64e8229d08c40ea58fbd7a2576d8559ba888c5 not found: ID does not exist" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.774449 4809 scope.go:117] "RemoveContainer" containerID="a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab" Nov 27 18:08:25 crc kubenswrapper[4809]: E1127 18:08:25.774894 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab\": container with ID starting with a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab not found: ID does not exist" containerID="a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.774933 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab"} err="failed to get container status \"a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab\": rpc error: code = NotFound desc = could not find container \"a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab\": container with ID starting with a4efb8cdc530a184f182befee4167cf6c97f6fa10fb85545ad68e87c1b6cfeab not found: ID does not exist" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.774974 4809 scope.go:117] "RemoveContainer" containerID="c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140" Nov 27 18:08:25 crc kubenswrapper[4809]: E1127 18:08:25.775290 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140\": container with ID starting with c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140 not found: ID does not exist" containerID="c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140" Nov 27 18:08:25 crc kubenswrapper[4809]: I1127 18:08:25.775344 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140"} err="failed to get container status \"c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140\": rpc error: code = NotFound desc = could not find container \"c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140\": container with ID starting with c0038bf69e3749de52cc6475d517f17657071ae3417fd34471b7233c67fbd140 not found: ID does not exist" Nov 27 18:08:26 crc kubenswrapper[4809]: I1127 18:08:26.014955 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kb7ck"] Nov 27 18:08:26 crc kubenswrapper[4809]: I1127 18:08:26.028386 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kb7ck"] Nov 27 18:08:27 crc kubenswrapper[4809]: I1127 18:08:27.470422 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" path="/var/lib/kubelet/pods/8d92e18f-7194-48c3-8db7-7f86b03289ec/volumes" Nov 27 18:08:55 crc kubenswrapper[4809]: I1127 18:08:55.779370 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:08:55 crc kubenswrapper[4809]: I1127 18:08:55.780058 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:08:55 crc kubenswrapper[4809]: I1127 18:08:55.993052 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.183:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 18:08:57 crc kubenswrapper[4809]: I1127 18:08:57.812874 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="56782df1-096c-47b7-bebb-b0a0712a46cc" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 27 18:09:25 crc kubenswrapper[4809]: I1127 18:09:25.779590 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:09:25 crc kubenswrapper[4809]: I1127 18:09:25.780178 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.475065 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f7msg"] Nov 27 18:09:38 crc kubenswrapper[4809]: E1127 18:09:38.476101 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerName="registry-server" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.476116 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerName="registry-server" Nov 27 18:09:38 crc kubenswrapper[4809]: E1127 18:09:38.476139 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerName="extract-utilities" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.476145 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerName="extract-utilities" Nov 27 18:09:38 crc kubenswrapper[4809]: E1127 18:09:38.476168 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerName="extract-content" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.476175 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerName="extract-content" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.476407 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d92e18f-7194-48c3-8db7-7f86b03289ec" containerName="registry-server" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.479560 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.489627 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7msg"] Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.576978 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-utilities\") pod \"redhat-marketplace-f7msg\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.577468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-catalog-content\") pod \"redhat-marketplace-f7msg\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.577720 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlbjk\" (UniqueName: \"kubernetes.io/projected/7f5913e3-2be0-4091-ba3b-492084683727-kube-api-access-vlbjk\") pod \"redhat-marketplace-f7msg\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.679556 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-catalog-content\") pod \"redhat-marketplace-f7msg\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.679651 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlbjk\" (UniqueName: \"kubernetes.io/projected/7f5913e3-2be0-4091-ba3b-492084683727-kube-api-access-vlbjk\") pod \"redhat-marketplace-f7msg\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.679783 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-utilities\") pod \"redhat-marketplace-f7msg\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.680143 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-catalog-content\") pod \"redhat-marketplace-f7msg\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.680250 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-utilities\") pod \"redhat-marketplace-f7msg\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.701255 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlbjk\" (UniqueName: \"kubernetes.io/projected/7f5913e3-2be0-4091-ba3b-492084683727-kube-api-access-vlbjk\") pod \"redhat-marketplace-f7msg\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:38 crc kubenswrapper[4809]: I1127 18:09:38.799056 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:39 crc kubenswrapper[4809]: I1127 18:09:39.260244 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7msg"] Nov 27 18:09:39 crc kubenswrapper[4809]: I1127 18:09:39.399115 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7msg" event={"ID":"7f5913e3-2be0-4091-ba3b-492084683727","Type":"ContainerStarted","Data":"6ec70e664d515f4c3eb77879648d2e8935b5c8f869f4c682bf6ba9dc77fec3df"} Nov 27 18:09:40 crc kubenswrapper[4809]: I1127 18:09:40.410641 4809 generic.go:334] "Generic (PLEG): container finished" podID="7f5913e3-2be0-4091-ba3b-492084683727" containerID="898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1" exitCode=0 Nov 27 18:09:40 crc kubenswrapper[4809]: I1127 18:09:40.410695 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7msg" event={"ID":"7f5913e3-2be0-4091-ba3b-492084683727","Type":"ContainerDied","Data":"898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1"} Nov 27 18:09:41 crc kubenswrapper[4809]: I1127 18:09:41.424171 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7msg" event={"ID":"7f5913e3-2be0-4091-ba3b-492084683727","Type":"ContainerStarted","Data":"24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c"} Nov 27 18:09:42 crc kubenswrapper[4809]: I1127 18:09:42.435383 4809 generic.go:334] "Generic (PLEG): container finished" podID="7f5913e3-2be0-4091-ba3b-492084683727" containerID="24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c" exitCode=0 Nov 27 18:09:42 crc kubenswrapper[4809]: I1127 18:09:42.435437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7msg" event={"ID":"7f5913e3-2be0-4091-ba3b-492084683727","Type":"ContainerDied","Data":"24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c"} Nov 27 18:09:43 crc kubenswrapper[4809]: I1127 18:09:43.448502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7msg" event={"ID":"7f5913e3-2be0-4091-ba3b-492084683727","Type":"ContainerStarted","Data":"d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2"} Nov 27 18:09:43 crc kubenswrapper[4809]: I1127 18:09:43.483012 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f7msg" podStartSLOduration=3.048191838 podStartE2EDuration="5.482992991s" podCreationTimestamp="2025-11-27 18:09:38 +0000 UTC" firstStartedPulling="2025-11-27 18:09:40.412624762 +0000 UTC m=+3615.685082114" lastFinishedPulling="2025-11-27 18:09:42.847425915 +0000 UTC m=+3618.119883267" observedRunningTime="2025-11-27 18:09:43.472473164 +0000 UTC m=+3618.744930516" watchObservedRunningTime="2025-11-27 18:09:43.482992991 +0000 UTC m=+3618.755450343" Nov 27 18:09:48 crc kubenswrapper[4809]: I1127 18:09:48.799640 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:48 crc kubenswrapper[4809]: I1127 18:09:48.800206 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:48 crc kubenswrapper[4809]: I1127 18:09:48.850931 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:49 crc kubenswrapper[4809]: I1127 18:09:49.556752 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:49 crc kubenswrapper[4809]: I1127 18:09:49.607878 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7msg"] Nov 27 18:09:51 crc kubenswrapper[4809]: I1127 18:09:51.529949 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f7msg" podUID="7f5913e3-2be0-4091-ba3b-492084683727" containerName="registry-server" containerID="cri-o://d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2" gracePeriod=2 Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.188973 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.260041 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-catalog-content\") pod \"7f5913e3-2be0-4091-ba3b-492084683727\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.260205 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlbjk\" (UniqueName: \"kubernetes.io/projected/7f5913e3-2be0-4091-ba3b-492084683727-kube-api-access-vlbjk\") pod \"7f5913e3-2be0-4091-ba3b-492084683727\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.260333 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-utilities\") pod \"7f5913e3-2be0-4091-ba3b-492084683727\" (UID: \"7f5913e3-2be0-4091-ba3b-492084683727\") " Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.261931 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-utilities" (OuterVolumeSpecName: "utilities") pod "7f5913e3-2be0-4091-ba3b-492084683727" (UID: "7f5913e3-2be0-4091-ba3b-492084683727"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.270140 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f5913e3-2be0-4091-ba3b-492084683727-kube-api-access-vlbjk" (OuterVolumeSpecName: "kube-api-access-vlbjk") pod "7f5913e3-2be0-4091-ba3b-492084683727" (UID: "7f5913e3-2be0-4091-ba3b-492084683727"). InnerVolumeSpecName "kube-api-access-vlbjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.290220 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f5913e3-2be0-4091-ba3b-492084683727" (UID: "7f5913e3-2be0-4091-ba3b-492084683727"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.362319 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.362367 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f5913e3-2be0-4091-ba3b-492084683727-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.362383 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlbjk\" (UniqueName: \"kubernetes.io/projected/7f5913e3-2be0-4091-ba3b-492084683727-kube-api-access-vlbjk\") on node \"crc\" DevicePath \"\"" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.542360 4809 generic.go:334] "Generic (PLEG): container finished" podID="7f5913e3-2be0-4091-ba3b-492084683727" containerID="d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2" exitCode=0 Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.542405 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7msg" event={"ID":"7f5913e3-2be0-4091-ba3b-492084683727","Type":"ContainerDied","Data":"d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2"} Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.542436 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7msg" event={"ID":"7f5913e3-2be0-4091-ba3b-492084683727","Type":"ContainerDied","Data":"6ec70e664d515f4c3eb77879648d2e8935b5c8f869f4c682bf6ba9dc77fec3df"} Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.542456 4809 scope.go:117] "RemoveContainer" containerID="d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.543155 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7msg" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.571951 4809 scope.go:117] "RemoveContainer" containerID="24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.581803 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7msg"] Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.589550 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7msg"] Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.601720 4809 scope.go:117] "RemoveContainer" containerID="898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.637760 4809 scope.go:117] "RemoveContainer" containerID="d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2" Nov 27 18:09:52 crc kubenswrapper[4809]: E1127 18:09:52.638168 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2\": container with ID starting with d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2 not found: ID does not exist" containerID="d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.638196 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2"} err="failed to get container status \"d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2\": rpc error: code = NotFound desc = could not find container \"d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2\": container with ID starting with d49361c520e94191dc84847775a42c3f365d1eb80ab08884b28d7e334ca07cd2 not found: ID does not exist" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.638216 4809 scope.go:117] "RemoveContainer" containerID="24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c" Nov 27 18:09:52 crc kubenswrapper[4809]: E1127 18:09:52.638587 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c\": container with ID starting with 24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c not found: ID does not exist" containerID="24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.638610 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c"} err="failed to get container status \"24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c\": rpc error: code = NotFound desc = could not find container \"24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c\": container with ID starting with 24485359a655f157bf9b0b8e949f5697e2fbf55010b6d62f415e15b29e9eda6c not found: ID does not exist" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.638624 4809 scope.go:117] "RemoveContainer" containerID="898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1" Nov 27 18:09:52 crc kubenswrapper[4809]: E1127 18:09:52.638866 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1\": container with ID starting with 898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1 not found: ID does not exist" containerID="898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1" Nov 27 18:09:52 crc kubenswrapper[4809]: I1127 18:09:52.638904 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1"} err="failed to get container status \"898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1\": rpc error: code = NotFound desc = could not find container \"898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1\": container with ID starting with 898409e2403ac9692964c4dff8e29d9d74088f8e31c8a98711c55089020acee1 not found: ID does not exist" Nov 27 18:09:53 crc kubenswrapper[4809]: I1127 18:09:53.483986 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f5913e3-2be0-4091-ba3b-492084683727" path="/var/lib/kubelet/pods/7f5913e3-2be0-4091-ba3b-492084683727/volumes" Nov 27 18:09:55 crc kubenswrapper[4809]: I1127 18:09:55.779003 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:09:55 crc kubenswrapper[4809]: I1127 18:09:55.779347 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:09:55 crc kubenswrapper[4809]: I1127 18:09:55.779391 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 18:09:55 crc kubenswrapper[4809]: I1127 18:09:55.780168 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb0a07bf48201ad85b3541400956b6a851a810403563fcd255ca613a34ae9a04"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 18:09:55 crc kubenswrapper[4809]: I1127 18:09:55.780222 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://fb0a07bf48201ad85b3541400956b6a851a810403563fcd255ca613a34ae9a04" gracePeriod=600 Nov 27 18:09:56 crc kubenswrapper[4809]: I1127 18:09:56.582281 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="fb0a07bf48201ad85b3541400956b6a851a810403563fcd255ca613a34ae9a04" exitCode=0 Nov 27 18:09:56 crc kubenswrapper[4809]: I1127 18:09:56.582352 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"fb0a07bf48201ad85b3541400956b6a851a810403563fcd255ca613a34ae9a04"} Nov 27 18:09:56 crc kubenswrapper[4809]: I1127 18:09:56.583159 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5"} Nov 27 18:09:56 crc kubenswrapper[4809]: I1127 18:09:56.583201 4809 scope.go:117] "RemoveContainer" containerID="2e9c67320da65ffd152a647982b5ba31b5e5d6428a6235687c8beaf7d4d09064" Nov 27 18:11:24 crc kubenswrapper[4809]: I1127 18:11:24.877162 4809 generic.go:334] "Generic (PLEG): container finished" podID="79f6ceba-551a-427c-8690-1c8db833367d" containerID="a156fa4fde29499a948a0620b7ef79f077f40346ad257a85ca2465823885e5b3" exitCode=0 Nov 27 18:11:24 crc kubenswrapper[4809]: I1127 18:11:24.877258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"79f6ceba-551a-427c-8690-1c8db833367d","Type":"ContainerDied","Data":"a156fa4fde29499a948a0620b7ef79f077f40346ad257a85ca2465823885e5b3"} Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.323003 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.418417 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ssh-key\") pod \"79f6ceba-551a-427c-8690-1c8db833367d\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.419202 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qxsl\" (UniqueName: \"kubernetes.io/projected/79f6ceba-551a-427c-8690-1c8db833367d-kube-api-access-8qxsl\") pod \"79f6ceba-551a-427c-8690-1c8db833367d\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.419271 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-workdir\") pod \"79f6ceba-551a-427c-8690-1c8db833367d\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.419303 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-config-data\") pod \"79f6ceba-551a-427c-8690-1c8db833367d\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.419327 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"79f6ceba-551a-427c-8690-1c8db833367d\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.419362 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ca-certs\") pod \"79f6ceba-551a-427c-8690-1c8db833367d\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.419394 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-temporary\") pod \"79f6ceba-551a-427c-8690-1c8db833367d\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.419424 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config-secret\") pod \"79f6ceba-551a-427c-8690-1c8db833367d\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.419525 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config\") pod \"79f6ceba-551a-427c-8690-1c8db833367d\" (UID: \"79f6ceba-551a-427c-8690-1c8db833367d\") " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.420641 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "79f6ceba-551a-427c-8690-1c8db833367d" (UID: "79f6ceba-551a-427c-8690-1c8db833367d"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.420772 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-config-data" (OuterVolumeSpecName: "config-data") pod "79f6ceba-551a-427c-8690-1c8db833367d" (UID: "79f6ceba-551a-427c-8690-1c8db833367d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.423068 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "79f6ceba-551a-427c-8690-1c8db833367d" (UID: "79f6ceba-551a-427c-8690-1c8db833367d"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.425139 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79f6ceba-551a-427c-8690-1c8db833367d-kube-api-access-8qxsl" (OuterVolumeSpecName: "kube-api-access-8qxsl") pod "79f6ceba-551a-427c-8690-1c8db833367d" (UID: "79f6ceba-551a-427c-8690-1c8db833367d"). InnerVolumeSpecName "kube-api-access-8qxsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.432832 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "79f6ceba-551a-427c-8690-1c8db833367d" (UID: "79f6ceba-551a-427c-8690-1c8db833367d"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.456871 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "79f6ceba-551a-427c-8690-1c8db833367d" (UID: "79f6ceba-551a-427c-8690-1c8db833367d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.460123 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "79f6ceba-551a-427c-8690-1c8db833367d" (UID: "79f6ceba-551a-427c-8690-1c8db833367d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.464894 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "79f6ceba-551a-427c-8690-1c8db833367d" (UID: "79f6ceba-551a-427c-8690-1c8db833367d"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.496054 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "79f6ceba-551a-427c-8690-1c8db833367d" (UID: "79f6ceba-551a-427c-8690-1c8db833367d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.522093 4809 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.522127 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.522141 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.522151 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.522161 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qxsl\" (UniqueName: \"kubernetes.io/projected/79f6ceba-551a-427c-8690-1c8db833367d-kube-api-access-8qxsl\") on node \"crc\" DevicePath \"\"" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.522169 4809 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/79f6ceba-551a-427c-8690-1c8db833367d-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.522178 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79f6ceba-551a-427c-8690-1c8db833367d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.522217 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.522228 4809 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/79f6ceba-551a-427c-8690-1c8db833367d-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.547331 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.624804 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.900646 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"79f6ceba-551a-427c-8690-1c8db833367d","Type":"ContainerDied","Data":"969d03f886578c0ca962b37aa51fc829faf08801b1a2a0c6b89003f4044a51b3"} Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.900717 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="969d03f886578c0ca962b37aa51fc829faf08801b1a2a0c6b89003f4044a51b3" Nov 27 18:11:27 crc kubenswrapper[4809]: I1127 18:11:26.900729 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.854487 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 27 18:11:30 crc kubenswrapper[4809]: E1127 18:11:30.855361 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5913e3-2be0-4091-ba3b-492084683727" containerName="extract-utilities" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.855376 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5913e3-2be0-4091-ba3b-492084683727" containerName="extract-utilities" Nov 27 18:11:30 crc kubenswrapper[4809]: E1127 18:11:30.855395 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5913e3-2be0-4091-ba3b-492084683727" containerName="extract-content" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.855403 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5913e3-2be0-4091-ba3b-492084683727" containerName="extract-content" Nov 27 18:11:30 crc kubenswrapper[4809]: E1127 18:11:30.855412 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5913e3-2be0-4091-ba3b-492084683727" containerName="registry-server" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.855418 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5913e3-2be0-4091-ba3b-492084683727" containerName="registry-server" Nov 27 18:11:30 crc kubenswrapper[4809]: E1127 18:11:30.855431 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f6ceba-551a-427c-8690-1c8db833367d" containerName="tempest-tests-tempest-tests-runner" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.855437 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f6ceba-551a-427c-8690-1c8db833367d" containerName="tempest-tests-tempest-tests-runner" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.855675 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f6ceba-551a-427c-8690-1c8db833367d" containerName="tempest-tests-tempest-tests-runner" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.855689 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f5913e3-2be0-4091-ba3b-492084683727" containerName="registry-server" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.856335 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.858453 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4g7qk" Nov 27 18:11:30 crc kubenswrapper[4809]: I1127 18:11:30.924307 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.013517 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3526155-0e52-4f2c-b07a-920b8cf61639\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.013567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgm55\" (UniqueName: \"kubernetes.io/projected/a3526155-0e52-4f2c-b07a-920b8cf61639-kube-api-access-bgm55\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3526155-0e52-4f2c-b07a-920b8cf61639\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.114917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3526155-0e52-4f2c-b07a-920b8cf61639\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.114964 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgm55\" (UniqueName: \"kubernetes.io/projected/a3526155-0e52-4f2c-b07a-920b8cf61639-kube-api-access-bgm55\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3526155-0e52-4f2c-b07a-920b8cf61639\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.115401 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3526155-0e52-4f2c-b07a-920b8cf61639\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.133029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgm55\" (UniqueName: \"kubernetes.io/projected/a3526155-0e52-4f2c-b07a-920b8cf61639-kube-api-access-bgm55\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3526155-0e52-4f2c-b07a-920b8cf61639\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.140860 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3526155-0e52-4f2c-b07a-920b8cf61639\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.175183 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.634572 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.636176 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 18:11:31 crc kubenswrapper[4809]: I1127 18:11:31.951733 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"a3526155-0e52-4f2c-b07a-920b8cf61639","Type":"ContainerStarted","Data":"1090dbf50a3ac1c190e3424d868cee0f856a9e9fa036e2e5b205dbf47de33fde"} Nov 27 18:11:32 crc kubenswrapper[4809]: I1127 18:11:32.963204 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"a3526155-0e52-4f2c-b07a-920b8cf61639","Type":"ContainerStarted","Data":"43e9d03acf38e869e5273bb70fce93930e4798c81c8d14c7a9eedd51744f1f7e"} Nov 27 18:11:32 crc kubenswrapper[4809]: I1127 18:11:32.983552 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.256998213 podStartE2EDuration="2.983532962s" podCreationTimestamp="2025-11-27 18:11:30 +0000 UTC" firstStartedPulling="2025-11-27 18:11:31.635613004 +0000 UTC m=+3726.908070356" lastFinishedPulling="2025-11-27 18:11:32.362147753 +0000 UTC m=+3727.634605105" observedRunningTime="2025-11-27 18:11:32.973970142 +0000 UTC m=+3728.246427494" watchObservedRunningTime="2025-11-27 18:11:32.983532962 +0000 UTC m=+3728.255990314" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.221728 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9khf6/must-gather-p5rhm"] Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.224365 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.226680 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-9khf6"/"default-dockercfg-cvgls" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.226949 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9khf6"/"openshift-service-ca.crt" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.226731 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9khf6"/"kube-root-ca.crt" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.244730 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9khf6/must-gather-p5rhm"] Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.354655 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4dqz\" (UniqueName: \"kubernetes.io/projected/fa97813a-2b2c-43a3-9084-06f4ae660067-kube-api-access-t4dqz\") pod \"must-gather-p5rhm\" (UID: \"fa97813a-2b2c-43a3-9084-06f4ae660067\") " pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.354901 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa97813a-2b2c-43a3-9084-06f4ae660067-must-gather-output\") pod \"must-gather-p5rhm\" (UID: \"fa97813a-2b2c-43a3-9084-06f4ae660067\") " pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.457032 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa97813a-2b2c-43a3-9084-06f4ae660067-must-gather-output\") pod \"must-gather-p5rhm\" (UID: \"fa97813a-2b2c-43a3-9084-06f4ae660067\") " pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.457504 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4dqz\" (UniqueName: \"kubernetes.io/projected/fa97813a-2b2c-43a3-9084-06f4ae660067-kube-api-access-t4dqz\") pod \"must-gather-p5rhm\" (UID: \"fa97813a-2b2c-43a3-9084-06f4ae660067\") " pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.457555 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa97813a-2b2c-43a3-9084-06f4ae660067-must-gather-output\") pod \"must-gather-p5rhm\" (UID: \"fa97813a-2b2c-43a3-9084-06f4ae660067\") " pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.477426 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4dqz\" (UniqueName: \"kubernetes.io/projected/fa97813a-2b2c-43a3-9084-06f4ae660067-kube-api-access-t4dqz\") pod \"must-gather-p5rhm\" (UID: \"fa97813a-2b2c-43a3-9084-06f4ae660067\") " pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:12:01 crc kubenswrapper[4809]: I1127 18:12:01.548252 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:12:02 crc kubenswrapper[4809]: I1127 18:12:02.013406 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9khf6/must-gather-p5rhm"] Nov 27 18:12:02 crc kubenswrapper[4809]: I1127 18:12:02.246465 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/must-gather-p5rhm" event={"ID":"fa97813a-2b2c-43a3-9084-06f4ae660067","Type":"ContainerStarted","Data":"94df3e54eb4714b70fb915090f5698b2ed635e41e6ce49580dc6a829dc08ec94"} Nov 27 18:12:08 crc kubenswrapper[4809]: I1127 18:12:08.312012 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/must-gather-p5rhm" event={"ID":"fa97813a-2b2c-43a3-9084-06f4ae660067","Type":"ContainerStarted","Data":"90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60"} Nov 27 18:12:09 crc kubenswrapper[4809]: I1127 18:12:09.323127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/must-gather-p5rhm" event={"ID":"fa97813a-2b2c-43a3-9084-06f4ae660067","Type":"ContainerStarted","Data":"dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1"} Nov 27 18:12:09 crc kubenswrapper[4809]: I1127 18:12:09.339676 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9khf6/must-gather-p5rhm" podStartSLOduration=2.475413611 podStartE2EDuration="8.339656922s" podCreationTimestamp="2025-11-27 18:12:01 +0000 UTC" firstStartedPulling="2025-11-27 18:12:02.016894123 +0000 UTC m=+3757.289351475" lastFinishedPulling="2025-11-27 18:12:07.881137434 +0000 UTC m=+3763.153594786" observedRunningTime="2025-11-27 18:12:09.335699614 +0000 UTC m=+3764.608156966" watchObservedRunningTime="2025-11-27 18:12:09.339656922 +0000 UTC m=+3764.612114274" Nov 27 18:12:12 crc kubenswrapper[4809]: I1127 18:12:12.093202 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9khf6/crc-debug-w98cs"] Nov 27 18:12:12 crc kubenswrapper[4809]: I1127 18:12:12.096135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:12:12 crc kubenswrapper[4809]: I1127 18:12:12.212727 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh22t\" (UniqueName: \"kubernetes.io/projected/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-kube-api-access-mh22t\") pod \"crc-debug-w98cs\" (UID: \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\") " pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:12:12 crc kubenswrapper[4809]: I1127 18:12:12.213169 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-host\") pod \"crc-debug-w98cs\" (UID: \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\") " pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:12:12 crc kubenswrapper[4809]: I1127 18:12:12.314470 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh22t\" (UniqueName: \"kubernetes.io/projected/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-kube-api-access-mh22t\") pod \"crc-debug-w98cs\" (UID: \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\") " pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:12:12 crc kubenswrapper[4809]: I1127 18:12:12.314559 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-host\") pod \"crc-debug-w98cs\" (UID: \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\") " pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:12:12 crc kubenswrapper[4809]: I1127 18:12:12.314805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-host\") pod \"crc-debug-w98cs\" (UID: \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\") " pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:12:12 crc kubenswrapper[4809]: I1127 18:12:12.334713 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh22t\" (UniqueName: \"kubernetes.io/projected/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-kube-api-access-mh22t\") pod \"crc-debug-w98cs\" (UID: \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\") " pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:12:12 crc kubenswrapper[4809]: I1127 18:12:12.418102 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:12:12 crc kubenswrapper[4809]: W1127 18:12:12.472268 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb11e5e04_a6a6_4a97_a3fd_8e58dc3a6cbf.slice/crio-48c8e73d9e4e2189a02405f8c83b0826a3362f2e8d2ae85a5cb922896badfb74 WatchSource:0}: Error finding container 48c8e73d9e4e2189a02405f8c83b0826a3362f2e8d2ae85a5cb922896badfb74: Status 404 returned error can't find the container with id 48c8e73d9e4e2189a02405f8c83b0826a3362f2e8d2ae85a5cb922896badfb74 Nov 27 18:12:13 crc kubenswrapper[4809]: I1127 18:12:13.358817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/crc-debug-w98cs" event={"ID":"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf","Type":"ContainerStarted","Data":"48c8e73d9e4e2189a02405f8c83b0826a3362f2e8d2ae85a5cb922896badfb74"} Nov 27 18:12:23 crc kubenswrapper[4809]: I1127 18:12:23.472245 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/crc-debug-w98cs" event={"ID":"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf","Type":"ContainerStarted","Data":"e0c6f745f8be6f6c52490ccc34098d713ad93d845d1b0ae2fc0cb5e9f4af389c"} Nov 27 18:12:23 crc kubenswrapper[4809]: I1127 18:12:23.489549 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9khf6/crc-debug-w98cs" podStartSLOduration=1.427623873 podStartE2EDuration="11.489532171s" podCreationTimestamp="2025-11-27 18:12:12 +0000 UTC" firstStartedPulling="2025-11-27 18:12:12.476543506 +0000 UTC m=+3767.749000858" lastFinishedPulling="2025-11-27 18:12:22.538451804 +0000 UTC m=+3777.810909156" observedRunningTime="2025-11-27 18:12:23.486866238 +0000 UTC m=+3778.759323590" watchObservedRunningTime="2025-11-27 18:12:23.489532171 +0000 UTC m=+3778.761989523" Nov 27 18:12:25 crc kubenswrapper[4809]: I1127 18:12:25.780180 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:12:25 crc kubenswrapper[4809]: I1127 18:12:25.780867 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:12:55 crc kubenswrapper[4809]: I1127 18:12:55.779937 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:12:55 crc kubenswrapper[4809]: I1127 18:12:55.780491 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:13:05 crc kubenswrapper[4809]: I1127 18:13:05.842670 4809 generic.go:334] "Generic (PLEG): container finished" podID="b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf" containerID="e0c6f745f8be6f6c52490ccc34098d713ad93d845d1b0ae2fc0cb5e9f4af389c" exitCode=0 Nov 27 18:13:05 crc kubenswrapper[4809]: I1127 18:13:05.842773 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/crc-debug-w98cs" event={"ID":"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf","Type":"ContainerDied","Data":"e0c6f745f8be6f6c52490ccc34098d713ad93d845d1b0ae2fc0cb5e9f4af389c"} Nov 27 18:13:06 crc kubenswrapper[4809]: I1127 18:13:06.972429 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.002703 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9khf6/crc-debug-w98cs"] Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.010670 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9khf6/crc-debug-w98cs"] Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.054901 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-host\") pod \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\" (UID: \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\") " Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.054988 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh22t\" (UniqueName: \"kubernetes.io/projected/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-kube-api-access-mh22t\") pod \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\" (UID: \"b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf\") " Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.054999 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-host" (OuterVolumeSpecName: "host") pod "b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf" (UID: "b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.055525 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-host\") on node \"crc\" DevicePath \"\"" Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.060833 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-kube-api-access-mh22t" (OuterVolumeSpecName: "kube-api-access-mh22t") pod "b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf" (UID: "b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf"). InnerVolumeSpecName "kube-api-access-mh22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.157540 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh22t\" (UniqueName: \"kubernetes.io/projected/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf-kube-api-access-mh22t\") on node \"crc\" DevicePath \"\"" Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.469602 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf" path="/var/lib/kubelet/pods/b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf/volumes" Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.863200 4809 scope.go:117] "RemoveContainer" containerID="e0c6f745f8be6f6c52490ccc34098d713ad93d845d1b0ae2fc0cb5e9f4af389c" Nov 27 18:13:07 crc kubenswrapper[4809]: I1127 18:13:07.863226 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-w98cs" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.161520 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9khf6/crc-debug-7f48w"] Nov 27 18:13:08 crc kubenswrapper[4809]: E1127 18:13:08.162448 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf" containerName="container-00" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.162513 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf" containerName="container-00" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.162756 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11e5e04-a6a6-4a97-a3fd-8e58dc3a6cbf" containerName="container-00" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.163465 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.276654 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eecdad67-ab0c-4395-b0c3-14f02998c0d9-host\") pod \"crc-debug-7f48w\" (UID: \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\") " pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.276756 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvtft\" (UniqueName: \"kubernetes.io/projected/eecdad67-ab0c-4395-b0c3-14f02998c0d9-kube-api-access-xvtft\") pod \"crc-debug-7f48w\" (UID: \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\") " pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.379115 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eecdad67-ab0c-4395-b0c3-14f02998c0d9-host\") pod \"crc-debug-7f48w\" (UID: \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\") " pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.379180 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvtft\" (UniqueName: \"kubernetes.io/projected/eecdad67-ab0c-4395-b0c3-14f02998c0d9-kube-api-access-xvtft\") pod \"crc-debug-7f48w\" (UID: \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\") " pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.379623 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eecdad67-ab0c-4395-b0c3-14f02998c0d9-host\") pod \"crc-debug-7f48w\" (UID: \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\") " pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.397173 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvtft\" (UniqueName: \"kubernetes.io/projected/eecdad67-ab0c-4395-b0c3-14f02998c0d9-kube-api-access-xvtft\") pod \"crc-debug-7f48w\" (UID: \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\") " pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.480765 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.873992 4809 generic.go:334] "Generic (PLEG): container finished" podID="eecdad67-ab0c-4395-b0c3-14f02998c0d9" containerID="2099aad695c2e550e3c372c3f8d3c82f3d199edee6002343567586ee536c352e" exitCode=0 Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.874089 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/crc-debug-7f48w" event={"ID":"eecdad67-ab0c-4395-b0c3-14f02998c0d9","Type":"ContainerDied","Data":"2099aad695c2e550e3c372c3f8d3c82f3d199edee6002343567586ee536c352e"} Nov 27 18:13:08 crc kubenswrapper[4809]: I1127 18:13:08.874369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/crc-debug-7f48w" event={"ID":"eecdad67-ab0c-4395-b0c3-14f02998c0d9","Type":"ContainerStarted","Data":"4698b65f5bd2158de8ca093c8f3504c79ddc8255e4406ed8653dd85240547d91"} Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.065118 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.131110 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9khf6/crc-debug-7f48w"] Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.139940 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9khf6/crc-debug-7f48w"] Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.222125 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvtft\" (UniqueName: \"kubernetes.io/projected/eecdad67-ab0c-4395-b0c3-14f02998c0d9-kube-api-access-xvtft\") pod \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\" (UID: \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\") " Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.222180 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eecdad67-ab0c-4395-b0c3-14f02998c0d9-host\") pod \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\" (UID: \"eecdad67-ab0c-4395-b0c3-14f02998c0d9\") " Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.222342 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eecdad67-ab0c-4395-b0c3-14f02998c0d9-host" (OuterVolumeSpecName: "host") pod "eecdad67-ab0c-4395-b0c3-14f02998c0d9" (UID: "eecdad67-ab0c-4395-b0c3-14f02998c0d9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.222872 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eecdad67-ab0c-4395-b0c3-14f02998c0d9-host\") on node \"crc\" DevicePath \"\"" Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.226613 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eecdad67-ab0c-4395-b0c3-14f02998c0d9-kube-api-access-xvtft" (OuterVolumeSpecName: "kube-api-access-xvtft") pod "eecdad67-ab0c-4395-b0c3-14f02998c0d9" (UID: "eecdad67-ab0c-4395-b0c3-14f02998c0d9"). InnerVolumeSpecName "kube-api-access-xvtft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.324593 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvtft\" (UniqueName: \"kubernetes.io/projected/eecdad67-ab0c-4395-b0c3-14f02998c0d9-kube-api-access-xvtft\") on node \"crc\" DevicePath \"\"" Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.907387 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4698b65f5bd2158de8ca093c8f3504c79ddc8255e4406ed8653dd85240547d91" Nov 27 18:13:10 crc kubenswrapper[4809]: I1127 18:13:10.907442 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-7f48w" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.285121 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9khf6/crc-debug-nnnqw"] Nov 27 18:13:11 crc kubenswrapper[4809]: E1127 18:13:11.285678 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eecdad67-ab0c-4395-b0c3-14f02998c0d9" containerName="container-00" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.285692 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="eecdad67-ab0c-4395-b0c3-14f02998c0d9" containerName="container-00" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.285950 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="eecdad67-ab0c-4395-b0c3-14f02998c0d9" containerName="container-00" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.286591 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.446502 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9slm\" (UniqueName: \"kubernetes.io/projected/a4692b36-dcef-4304-997f-b11aa1a6b12b-kube-api-access-s9slm\") pod \"crc-debug-nnnqw\" (UID: \"a4692b36-dcef-4304-997f-b11aa1a6b12b\") " pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.446710 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a4692b36-dcef-4304-997f-b11aa1a6b12b-host\") pod \"crc-debug-nnnqw\" (UID: \"a4692b36-dcef-4304-997f-b11aa1a6b12b\") " pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.470014 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eecdad67-ab0c-4395-b0c3-14f02998c0d9" path="/var/lib/kubelet/pods/eecdad67-ab0c-4395-b0c3-14f02998c0d9/volumes" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.548482 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9slm\" (UniqueName: \"kubernetes.io/projected/a4692b36-dcef-4304-997f-b11aa1a6b12b-kube-api-access-s9slm\") pod \"crc-debug-nnnqw\" (UID: \"a4692b36-dcef-4304-997f-b11aa1a6b12b\") " pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.548672 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a4692b36-dcef-4304-997f-b11aa1a6b12b-host\") pod \"crc-debug-nnnqw\" (UID: \"a4692b36-dcef-4304-997f-b11aa1a6b12b\") " pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.548814 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a4692b36-dcef-4304-997f-b11aa1a6b12b-host\") pod \"crc-debug-nnnqw\" (UID: \"a4692b36-dcef-4304-997f-b11aa1a6b12b\") " pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.566644 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9slm\" (UniqueName: \"kubernetes.io/projected/a4692b36-dcef-4304-997f-b11aa1a6b12b-kube-api-access-s9slm\") pod \"crc-debug-nnnqw\" (UID: \"a4692b36-dcef-4304-997f-b11aa1a6b12b\") " pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.605641 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:11 crc kubenswrapper[4809]: W1127 18:13:11.629970 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4692b36_dcef_4304_997f_b11aa1a6b12b.slice/crio-46c4fc2480852e5783c4d32e937b7204002a0cecec09312daf460ceae3795403 WatchSource:0}: Error finding container 46c4fc2480852e5783c4d32e937b7204002a0cecec09312daf460ceae3795403: Status 404 returned error can't find the container with id 46c4fc2480852e5783c4d32e937b7204002a0cecec09312daf460ceae3795403 Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.918963 4809 generic.go:334] "Generic (PLEG): container finished" podID="a4692b36-dcef-4304-997f-b11aa1a6b12b" containerID="4fa2fd64877148b210e16857f45d149545c8f712cd22c37d4e452115fb2c941f" exitCode=0 Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.919055 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/crc-debug-nnnqw" event={"ID":"a4692b36-dcef-4304-997f-b11aa1a6b12b","Type":"ContainerDied","Data":"4fa2fd64877148b210e16857f45d149545c8f712cd22c37d4e452115fb2c941f"} Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.919542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/crc-debug-nnnqw" event={"ID":"a4692b36-dcef-4304-997f-b11aa1a6b12b","Type":"ContainerStarted","Data":"46c4fc2480852e5783c4d32e937b7204002a0cecec09312daf460ceae3795403"} Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.963195 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9khf6/crc-debug-nnnqw"] Nov 27 18:13:11 crc kubenswrapper[4809]: I1127 18:13:11.974020 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9khf6/crc-debug-nnnqw"] Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.055707 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.179468 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a4692b36-dcef-4304-997f-b11aa1a6b12b-host\") pod \"a4692b36-dcef-4304-997f-b11aa1a6b12b\" (UID: \"a4692b36-dcef-4304-997f-b11aa1a6b12b\") " Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.179628 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9slm\" (UniqueName: \"kubernetes.io/projected/a4692b36-dcef-4304-997f-b11aa1a6b12b-kube-api-access-s9slm\") pod \"a4692b36-dcef-4304-997f-b11aa1a6b12b\" (UID: \"a4692b36-dcef-4304-997f-b11aa1a6b12b\") " Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.179642 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4692b36-dcef-4304-997f-b11aa1a6b12b-host" (OuterVolumeSpecName: "host") pod "a4692b36-dcef-4304-997f-b11aa1a6b12b" (UID: "a4692b36-dcef-4304-997f-b11aa1a6b12b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.180173 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a4692b36-dcef-4304-997f-b11aa1a6b12b-host\") on node \"crc\" DevicePath \"\"" Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.185957 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4692b36-dcef-4304-997f-b11aa1a6b12b-kube-api-access-s9slm" (OuterVolumeSpecName: "kube-api-access-s9slm") pod "a4692b36-dcef-4304-997f-b11aa1a6b12b" (UID: "a4692b36-dcef-4304-997f-b11aa1a6b12b"). InnerVolumeSpecName "kube-api-access-s9slm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.282145 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9slm\" (UniqueName: \"kubernetes.io/projected/a4692b36-dcef-4304-997f-b11aa1a6b12b-kube-api-access-s9slm\") on node \"crc\" DevicePath \"\"" Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.470068 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4692b36-dcef-4304-997f-b11aa1a6b12b" path="/var/lib/kubelet/pods/a4692b36-dcef-4304-997f-b11aa1a6b12b/volumes" Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.939091 4809 scope.go:117] "RemoveContainer" containerID="4fa2fd64877148b210e16857f45d149545c8f712cd22c37d4e452115fb2c941f" Nov 27 18:13:13 crc kubenswrapper[4809]: I1127 18:13:13.939112 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/crc-debug-nnnqw" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.126021 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v4t67"] Nov 27 18:13:24 crc kubenswrapper[4809]: E1127 18:13:24.127004 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4692b36-dcef-4304-997f-b11aa1a6b12b" containerName="container-00" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.127019 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4692b36-dcef-4304-997f-b11aa1a6b12b" containerName="container-00" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.127260 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4692b36-dcef-4304-997f-b11aa1a6b12b" containerName="container-00" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.128724 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.146892 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v4t67"] Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.184447 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-catalog-content\") pod \"community-operators-v4t67\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.184557 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-utilities\") pod \"community-operators-v4t67\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.184597 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwlv9\" (UniqueName: \"kubernetes.io/projected/4ccebb66-4111-4960-ba0c-4510f5ce941b-kube-api-access-cwlv9\") pod \"community-operators-v4t67\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.286751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-utilities\") pod \"community-operators-v4t67\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.286811 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwlv9\" (UniqueName: \"kubernetes.io/projected/4ccebb66-4111-4960-ba0c-4510f5ce941b-kube-api-access-cwlv9\") pod \"community-operators-v4t67\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.286902 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-catalog-content\") pod \"community-operators-v4t67\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.287278 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-catalog-content\") pod \"community-operators-v4t67\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.288714 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-utilities\") pod \"community-operators-v4t67\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.305954 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwlv9\" (UniqueName: \"kubernetes.io/projected/4ccebb66-4111-4960-ba0c-4510f5ce941b-kube-api-access-cwlv9\") pod \"community-operators-v4t67\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.450467 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:24 crc kubenswrapper[4809]: W1127 18:13:24.891295 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ccebb66_4111_4960_ba0c_4510f5ce941b.slice/crio-4b984d93a108c11bd043a22706f4b95d749926d134724e4f8d37dabfea6851f0 WatchSource:0}: Error finding container 4b984d93a108c11bd043a22706f4b95d749926d134724e4f8d37dabfea6851f0: Status 404 returned error can't find the container with id 4b984d93a108c11bd043a22706f4b95d749926d134724e4f8d37dabfea6851f0 Nov 27 18:13:24 crc kubenswrapper[4809]: I1127 18:13:24.901079 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v4t67"] Nov 27 18:13:25 crc kubenswrapper[4809]: I1127 18:13:25.046581 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4t67" event={"ID":"4ccebb66-4111-4960-ba0c-4510f5ce941b","Type":"ContainerStarted","Data":"4b984d93a108c11bd043a22706f4b95d749926d134724e4f8d37dabfea6851f0"} Nov 27 18:13:25 crc kubenswrapper[4809]: I1127 18:13:25.779570 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:13:25 crc kubenswrapper[4809]: I1127 18:13:25.779987 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:13:25 crc kubenswrapper[4809]: I1127 18:13:25.780034 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 18:13:25 crc kubenswrapper[4809]: I1127 18:13:25.780775 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 18:13:25 crc kubenswrapper[4809]: I1127 18:13:25.780819 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" gracePeriod=600 Nov 27 18:13:25 crc kubenswrapper[4809]: E1127 18:13:25.898826 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:13:26 crc kubenswrapper[4809]: I1127 18:13:26.063696 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" exitCode=0 Nov 27 18:13:26 crc kubenswrapper[4809]: I1127 18:13:26.064084 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5"} Nov 27 18:13:26 crc kubenswrapper[4809]: I1127 18:13:26.064201 4809 scope.go:117] "RemoveContainer" containerID="fb0a07bf48201ad85b3541400956b6a851a810403563fcd255ca613a34ae9a04" Nov 27 18:13:26 crc kubenswrapper[4809]: I1127 18:13:26.064969 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:13:26 crc kubenswrapper[4809]: E1127 18:13:26.065595 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:13:26 crc kubenswrapper[4809]: I1127 18:13:26.069308 4809 generic.go:334] "Generic (PLEG): container finished" podID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerID="2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500" exitCode=0 Nov 27 18:13:26 crc kubenswrapper[4809]: I1127 18:13:26.069352 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4t67" event={"ID":"4ccebb66-4111-4960-ba0c-4510f5ce941b","Type":"ContainerDied","Data":"2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500"} Nov 27 18:13:27 crc kubenswrapper[4809]: I1127 18:13:27.345933 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5bfc74dc5d-snbcn_fea753ab-13fb-49aa-a430-6c5db50f1e6b/barbican-api/0.log" Nov 27 18:13:27 crc kubenswrapper[4809]: I1127 18:13:27.455464 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5bfc74dc5d-snbcn_fea753ab-13fb-49aa-a430-6c5db50f1e6b/barbican-api-log/0.log" Nov 27 18:13:27 crc kubenswrapper[4809]: I1127 18:13:27.618638 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-84d69cc6b6-vqlb7_9bc3b7a8-110d-481e-95d5-3d9034c0b2f9/barbican-keystone-listener/0.log" Nov 27 18:13:27 crc kubenswrapper[4809]: I1127 18:13:27.639455 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-84d69cc6b6-vqlb7_9bc3b7a8-110d-481e-95d5-3d9034c0b2f9/barbican-keystone-listener-log/0.log" Nov 27 18:13:27 crc kubenswrapper[4809]: I1127 18:13:27.721649 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-65b4758f4c-5krfp_af7083a1-8a26-49b0-8b7a-7b412592ed60/barbican-worker/0.log" Nov 27 18:13:27 crc kubenswrapper[4809]: I1127 18:13:27.808768 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-65b4758f4c-5krfp_af7083a1-8a26-49b0-8b7a-7b412592ed60/barbican-worker-log/0.log" Nov 27 18:13:27 crc kubenswrapper[4809]: I1127 18:13:27.941970 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq_1f21abe6-3443-4a72-b93b-574203c3f3b4/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.019670 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56782df1-096c-47b7-bebb-b0a0712a46cc/ceilometer-central-agent/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.073837 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56782df1-096c-47b7-bebb-b0a0712a46cc/ceilometer-notification-agent/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.091623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4t67" event={"ID":"4ccebb66-4111-4960-ba0c-4510f5ce941b","Type":"ContainerStarted","Data":"e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740"} Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.100889 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56782df1-096c-47b7-bebb-b0a0712a46cc/sg-core/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.139143 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56782df1-096c-47b7-bebb-b0a0712a46cc/proxy-httpd/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.282494 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6/cinder-api/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.297162 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6/cinder-api-log/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.477961 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92/probe/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.512038 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92/cinder-scheduler/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.546564 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-78ns4_2021472c-275a-44fd-802b-37e72374a72d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.707566 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb_0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.771070 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-zgq88_0025bfa8-25eb-43cc-825e-c4e299f28d47/init/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.894934 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-zgq88_0025bfa8-25eb-43cc-825e-c4e299f28d47/init/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.928855 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-zgq88_0025bfa8-25eb-43cc-825e-c4e299f28d47/dnsmasq-dns/0.log" Nov 27 18:13:28 crc kubenswrapper[4809]: I1127 18:13:28.958756 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj_e5b77adb-7f43-4df6-b676-cb82003887fc/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.101097 4809 generic.go:334] "Generic (PLEG): container finished" podID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerID="e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740" exitCode=0 Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.101136 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4t67" event={"ID":"4ccebb66-4111-4960-ba0c-4510f5ce941b","Type":"ContainerDied","Data":"e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740"} Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.105425 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2339b0f8-362a-43b7-8679-8259e295d65e/glance-httpd/0.log" Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.169077 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2339b0f8-362a-43b7-8679-8259e295d65e/glance-log/0.log" Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.307216 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_78ba6922-17ec-497d-b15d-e2ba250a698e/glance-log/0.log" Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.312589 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_78ba6922-17ec-497d-b15d-e2ba250a698e/glance-httpd/0.log" Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.396115 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr_43c7fb64-1bd5-4341-81e9-6129b4121106/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.507416 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-8wfrh_faa8dfb6-0171-4ad3-a68f-46ad6dd83646/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.784248 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29404441-hv6j5_47d4ccb6-b0c6-4676-a5d0-eb82e862457d/keystone-cron/0.log" Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.811880 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-856cbdcf9c-lws7x_5ddd7431-85e1-4f2e-ae95-f00162cbe120/keystone-api/0.log" Nov 27 18:13:29 crc kubenswrapper[4809]: I1127 18:13:29.943609 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3ff647e6-3fcb-401c-a695-047f07ca9a58/kube-state-metrics/0.log" Nov 27 18:13:30 crc kubenswrapper[4809]: I1127 18:13:30.088577 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp_80c21e8c-3128-48ef-91a8-365409103274/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:30 crc kubenswrapper[4809]: I1127 18:13:30.187729 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-ddw9s_565ebe47-4144-4730-9e10-ebd98010c9a3/logging-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:30 crc kubenswrapper[4809]: I1127 18:13:30.579421 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-65c6b6d897-rhmzw_1a7bab33-46e8-4e18-a0c2-1ca25e22493f/neutron-httpd/0.log" Nov 27 18:13:30 crc kubenswrapper[4809]: I1127 18:13:30.654703 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-65c6b6d897-rhmzw_1a7bab33-46e8-4e18-a0c2-1ca25e22493f/neutron-api/0.log" Nov 27 18:13:30 crc kubenswrapper[4809]: I1127 18:13:30.916959 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f_0d3439fb-cf21-47a9-bbf7-39a34a5469dd/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:31 crc kubenswrapper[4809]: I1127 18:13:31.120900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4t67" event={"ID":"4ccebb66-4111-4960-ba0c-4510f5ce941b","Type":"ContainerStarted","Data":"3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f"} Nov 27 18:13:31 crc kubenswrapper[4809]: I1127 18:13:31.152456 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v4t67" podStartSLOduration=2.7702422159999998 podStartE2EDuration="7.152430578s" podCreationTimestamp="2025-11-27 18:13:24 +0000 UTC" firstStartedPulling="2025-11-27 18:13:26.070979751 +0000 UTC m=+3841.343437103" lastFinishedPulling="2025-11-27 18:13:30.453168113 +0000 UTC m=+3845.725625465" observedRunningTime="2025-11-27 18:13:31.137837679 +0000 UTC m=+3846.410295051" watchObservedRunningTime="2025-11-27 18:13:31.152430578 +0000 UTC m=+3846.424887930" Nov 27 18:13:31 crc kubenswrapper[4809]: I1127 18:13:31.228099 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_db48eb93-0220-4661-8de6-23aa1e43ca94/nova-api-log/0.log" Nov 27 18:13:31 crc kubenswrapper[4809]: I1127 18:13:31.292457 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b/nova-cell0-conductor-conductor/0.log" Nov 27 18:13:31 crc kubenswrapper[4809]: I1127 18:13:31.434957 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_db48eb93-0220-4661-8de6-23aa1e43ca94/nova-api-api/0.log" Nov 27 18:13:31 crc kubenswrapper[4809]: I1127 18:13:31.591712 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_01ecca97-66c5-4670-9304-5fb2b116e4be/nova-cell1-conductor-conductor/0.log" Nov 27 18:13:31 crc kubenswrapper[4809]: I1127 18:13:31.642184 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_969b1f1d-ce0b-48fd-b149-72649f406cf0/nova-cell1-novncproxy-novncproxy/0.log" Nov 27 18:13:31 crc kubenswrapper[4809]: I1127 18:13:31.791666 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-zc795_9230a20e-bbed-4c09-8e94-a87279b78723/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:31 crc kubenswrapper[4809]: I1127 18:13:31.939278 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_65be15f8-d10f-4079-b79a-eb92d2cdc432/nova-metadata-log/0.log" Nov 27 18:13:32 crc kubenswrapper[4809]: I1127 18:13:32.240416 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f5bbe948-59f7-488a-b62f-ea8ead20541f/nova-scheduler-scheduler/0.log" Nov 27 18:13:32 crc kubenswrapper[4809]: I1127 18:13:32.277136 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f82a4054-fea3-4071-b65d-529d900c0f04/mysql-bootstrap/0.log" Nov 27 18:13:32 crc kubenswrapper[4809]: I1127 18:13:32.474007 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f82a4054-fea3-4071-b65d-529d900c0f04/mysql-bootstrap/0.log" Nov 27 18:13:32 crc kubenswrapper[4809]: I1127 18:13:32.483649 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f82a4054-fea3-4071-b65d-529d900c0f04/galera/0.log" Nov 27 18:13:32 crc kubenswrapper[4809]: I1127 18:13:32.709545 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3cf12fa-2068-4bcd-ae42-d7c6dfea6782/mysql-bootstrap/0.log" Nov 27 18:13:32 crc kubenswrapper[4809]: I1127 18:13:32.891526 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_65be15f8-d10f-4079-b79a-eb92d2cdc432/nova-metadata-metadata/0.log" Nov 27 18:13:32 crc kubenswrapper[4809]: I1127 18:13:32.927909 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3cf12fa-2068-4bcd-ae42-d7c6dfea6782/mysql-bootstrap/0.log" Nov 27 18:13:32 crc kubenswrapper[4809]: I1127 18:13:32.969642 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3cf12fa-2068-4bcd-ae42-d7c6dfea6782/galera/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.152081 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_090560a3-12be-4a67-b507-0c7764053bf2/openstackclient/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.235689 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-h8g6t_55d50ebb-8cab-42df-96b5-9598262337a4/ovn-controller/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.342105 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9567r_7a4f76ce-dff9-4d9c-a75b-7282bc630a36/openstack-network-exporter/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.484296 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jnvf9_a501e4bf-f184-451f-9b21-a813f147413d/ovsdb-server-init/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.668665 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jnvf9_a501e4bf-f184-451f-9b21-a813f147413d/ovs-vswitchd/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.669701 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jnvf9_a501e4bf-f184-451f-9b21-a813f147413d/ovsdb-server-init/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.692306 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jnvf9_a501e4bf-f184-451f-9b21-a813f147413d/ovsdb-server/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.891983 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-dwhp4_cc517a71-c16f-4144-94d8-36c2878d89d8/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.901380 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_545a6dbd-4f59-435b-b951-0a8e58be0d2a/openstack-network-exporter/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.906630 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_bb517228-deb6-41fa-b6aa-354014cd34e1/memcached/0.log" Nov 27 18:13:33 crc kubenswrapper[4809]: I1127 18:13:33.928835 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_545a6dbd-4f59-435b-b951-0a8e58be0d2a/ovn-northd/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.074770 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6ab5ce4-ce26-4eee-98fe-1bb380df9376/openstack-network-exporter/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.084328 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6ab5ce4-ce26-4eee-98fe-1bb380df9376/ovsdbserver-nb/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.120157 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cda8dbab-96fd-4d1b-a9cb-e056382df1a2/openstack-network-exporter/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.262630 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cda8dbab-96fd-4d1b-a9cb-e056382df1a2/ovsdbserver-sb/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.290279 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-556f88f66d-dk28p_a43816b3-bfa7-4570-b082-9b217d660416/placement-api/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.379214 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-556f88f66d-dk28p_a43816b3-bfa7-4570-b082-9b217d660416/placement-log/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.428325 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3f0cad7c-85ca-4269-be47-555613716799/setup-container/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.450619 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.450707 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.498003 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.618855 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3f0cad7c-85ca-4269-be47-555613716799/setup-container/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.626469 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3f0cad7c-85ca-4269-be47-555613716799/rabbitmq/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.647066 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_62edd287-129b-48b0-af61-d21bd27590d7/setup-container/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.821182 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_62edd287-129b-48b0-af61-d21bd27590d7/setup-container/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.851173 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_62edd287-129b-48b0-af61-d21bd27590d7/rabbitmq/0.log" Nov 27 18:13:34 crc kubenswrapper[4809]: I1127 18:13:34.883292 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn_562a0647-54b3-4b56-9717-24f2ab4ca87f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.020256 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-w5fcv_cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.061012 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm_56b0a916-b733-4a1c-b6dc-97925e319b83/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.118525 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-v55jx_5beb9875-497c-4eb8-9b9d-8474a6891b82/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.205838 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.263855 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v4t67"] Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.271368 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-wvz7x_7616f14b-3549-40b3-ba73-32b1cb830d98/ssh-known-hosts-edpm-deployment/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.389677 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5999bb96c5-jlmwd_2cdfac16-f905-422d-a059-60d8e3238bcf/proxy-server/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.421445 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5999bb96c5-jlmwd_2cdfac16-f905-422d-a059-60d8e3238bcf/proxy-httpd/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.510694 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-hgrfv_98c79518-5ef3-4382-a156-6de22fc92082/swift-ring-rebalance/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.584075 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/account-auditor/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.630309 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/account-reaper/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.652656 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/account-replicator/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.710215 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/account-server/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.766423 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/container-auditor/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.809488 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/container-replicator/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.814110 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/container-server/0.log" Nov 27 18:13:35 crc kubenswrapper[4809]: I1127 18:13:35.863774 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/container-updater/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.063150 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-auditor/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.083164 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-expirer/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.120348 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-replicator/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.175006 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-updater/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.178147 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-server/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.256321 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/swift-recon-cron/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.278370 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/rsync/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.382829 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kv495_10c284aa-b6df-41bb-8ace-7dd18b804925/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.461904 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_79f6ceba-551a-427c-8690-1c8db833367d/tempest-tests-tempest-tests-runner/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.578311 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_a3526155-0e52-4f2c-b07a-920b8cf61639/test-operator-logs-container/0.log" Nov 27 18:13:36 crc kubenswrapper[4809]: I1127 18:13:36.604889 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7_c91fe577-8e90-411e-b078-8ece235f2f93/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.170430 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v4t67" podUID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerName="registry-server" containerID="cri-o://3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f" gracePeriod=2 Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.690780 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.848190 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwlv9\" (UniqueName: \"kubernetes.io/projected/4ccebb66-4111-4960-ba0c-4510f5ce941b-kube-api-access-cwlv9\") pod \"4ccebb66-4111-4960-ba0c-4510f5ce941b\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.848324 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-utilities\") pod \"4ccebb66-4111-4960-ba0c-4510f5ce941b\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.848460 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-catalog-content\") pod \"4ccebb66-4111-4960-ba0c-4510f5ce941b\" (UID: \"4ccebb66-4111-4960-ba0c-4510f5ce941b\") " Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.849316 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-utilities" (OuterVolumeSpecName: "utilities") pod "4ccebb66-4111-4960-ba0c-4510f5ce941b" (UID: "4ccebb66-4111-4960-ba0c-4510f5ce941b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.854702 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ccebb66-4111-4960-ba0c-4510f5ce941b-kube-api-access-cwlv9" (OuterVolumeSpecName: "kube-api-access-cwlv9") pod "4ccebb66-4111-4960-ba0c-4510f5ce941b" (UID: "4ccebb66-4111-4960-ba0c-4510f5ce941b"). InnerVolumeSpecName "kube-api-access-cwlv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.911475 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ccebb66-4111-4960-ba0c-4510f5ce941b" (UID: "4ccebb66-4111-4960-ba0c-4510f5ce941b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.951454 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.951493 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ccebb66-4111-4960-ba0c-4510f5ce941b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:13:37 crc kubenswrapper[4809]: I1127 18:13:37.951503 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwlv9\" (UniqueName: \"kubernetes.io/projected/4ccebb66-4111-4960-ba0c-4510f5ce941b-kube-api-access-cwlv9\") on node \"crc\" DevicePath \"\"" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.179479 4809 generic.go:334] "Generic (PLEG): container finished" podID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerID="3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f" exitCode=0 Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.179521 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4t67" event={"ID":"4ccebb66-4111-4960-ba0c-4510f5ce941b","Type":"ContainerDied","Data":"3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f"} Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.179547 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4t67" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.179567 4809 scope.go:117] "RemoveContainer" containerID="3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.179553 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4t67" event={"ID":"4ccebb66-4111-4960-ba0c-4510f5ce941b","Type":"ContainerDied","Data":"4b984d93a108c11bd043a22706f4b95d749926d134724e4f8d37dabfea6851f0"} Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.197979 4809 scope.go:117] "RemoveContainer" containerID="e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.218097 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v4t67"] Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.225904 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v4t67"] Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.227016 4809 scope.go:117] "RemoveContainer" containerID="2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.268779 4809 scope.go:117] "RemoveContainer" containerID="3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f" Nov 27 18:13:38 crc kubenswrapper[4809]: E1127 18:13:38.269432 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f\": container with ID starting with 3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f not found: ID does not exist" containerID="3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.269463 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f"} err="failed to get container status \"3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f\": rpc error: code = NotFound desc = could not find container \"3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f\": container with ID starting with 3edc10a46e206ee0f4ab88fbbb0b0c368bfc00c0182b15f6be67ffeb5fbe801f not found: ID does not exist" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.269483 4809 scope.go:117] "RemoveContainer" containerID="e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740" Nov 27 18:13:38 crc kubenswrapper[4809]: E1127 18:13:38.269867 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740\": container with ID starting with e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740 not found: ID does not exist" containerID="e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.269906 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740"} err="failed to get container status \"e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740\": rpc error: code = NotFound desc = could not find container \"e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740\": container with ID starting with e58683a7764fba9350480817be5f5fb05c4661ca47373bd18fd2e78460363740 not found: ID does not exist" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.269961 4809 scope.go:117] "RemoveContainer" containerID="2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500" Nov 27 18:13:38 crc kubenswrapper[4809]: E1127 18:13:38.270249 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500\": container with ID starting with 2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500 not found: ID does not exist" containerID="2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500" Nov 27 18:13:38 crc kubenswrapper[4809]: I1127 18:13:38.270277 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500"} err="failed to get container status \"2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500\": rpc error: code = NotFound desc = could not find container \"2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500\": container with ID starting with 2adea1a4dc753454d99e0464622ea9ba31c8fc3b4b336a1a748e264294839500 not found: ID does not exist" Nov 27 18:13:39 crc kubenswrapper[4809]: I1127 18:13:39.472122 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ccebb66-4111-4960-ba0c-4510f5ce941b" path="/var/lib/kubelet/pods/4ccebb66-4111-4960-ba0c-4510f5ce941b/volumes" Nov 27 18:13:40 crc kubenswrapper[4809]: I1127 18:13:40.458465 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:13:40 crc kubenswrapper[4809]: E1127 18:13:40.459006 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:13:54 crc kubenswrapper[4809]: I1127 18:13:54.458374 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:13:54 crc kubenswrapper[4809]: E1127 18:13:54.459135 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:13:55 crc kubenswrapper[4809]: I1127 18:13:55.494384 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/util/0.log" Nov 27 18:13:55 crc kubenswrapper[4809]: I1127 18:13:55.667110 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/pull/0.log" Nov 27 18:13:55 crc kubenswrapper[4809]: I1127 18:13:55.670487 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/util/0.log" Nov 27 18:13:55 crc kubenswrapper[4809]: I1127 18:13:55.705255 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/pull/0.log" Nov 27 18:13:55 crc kubenswrapper[4809]: I1127 18:13:55.874029 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/util/0.log" Nov 27 18:13:55 crc kubenswrapper[4809]: I1127 18:13:55.882402 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/pull/0.log" Nov 27 18:13:55 crc kubenswrapper[4809]: I1127 18:13:55.915520 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/extract/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.044944 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-g7pm5_4e5fbc76-cd41-432e-8c5b-90ff77d0704c/kube-rbac-proxy/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.097890 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-g7pm5_4e5fbc76-cd41-432e-8c5b-90ff77d0704c/manager/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.111960 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-swx59_be316447-6b83-4563-b75e-d4088882b695/kube-rbac-proxy/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.259653 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-swx59_be316447-6b83-4563-b75e-d4088882b695/manager/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.294590 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-l58tv_b470d686-15fd-4d31-ac6a-ffb60d3c9d79/kube-rbac-proxy/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.299911 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-l58tv_b470d686-15fd-4d31-ac6a-ffb60d3c9d79/manager/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.468103 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-x4fdl_f6774dea-1e4a-4cd0-8688-5a588fe6d49b/kube-rbac-proxy/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.572464 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-x4fdl_f6774dea-1e4a-4cd0-8688-5a588fe6d49b/manager/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.608721 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-4rlvc_27f6291c-27aa-430f-864e-05be868c89b9/kube-rbac-proxy/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.693545 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-4rlvc_27f6291c-27aa-430f-864e-05be868c89b9/manager/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.763953 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-mkngh_16fc4440-2960-4b7d-b059-c5192ea37a50/kube-rbac-proxy/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.793268 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-mkngh_16fc4440-2960-4b7d-b059-c5192ea37a50/manager/0.log" Nov 27 18:13:56 crc kubenswrapper[4809]: I1127 18:13:56.940599 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-b28r8_aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1/kube-rbac-proxy/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.086402 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-b28r8_aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1/manager/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.093446 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-x9qq5_976fdd57-7e54-4896-abca-914b86a98b19/kube-rbac-proxy/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.124728 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-x9qq5_976fdd57-7e54-4896-abca-914b86a98b19/manager/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.257828 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-mwv49_63be6575-532b-487f-97a2-d9fd077e5be0/kube-rbac-proxy/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.345937 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-mwv49_63be6575-532b-487f-97a2-d9fd077e5be0/manager/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.440840 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-bzjd8_ad51c9a8-6543-403b-b8de-ff189da761a4/kube-rbac-proxy/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.444012 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-bzjd8_ad51c9a8-6543-403b-b8de-ff189da761a4/manager/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.542814 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-tcwh6_9fb9364a-e568-418b-b403-dfa071e60297/kube-rbac-proxy/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.640614 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-tcwh6_9fb9364a-e568-418b-b403-dfa071e60297/manager/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.701278 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-6t8sk_db1b7072-b244-4e32-ba19-b8cd4254af2f/kube-rbac-proxy/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.810046 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-ll5l2_da239d06-b1c7-4b06-b830-5108030beed6/kube-rbac-proxy/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.829536 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-6t8sk_db1b7072-b244-4e32-ba19-b8cd4254af2f/manager/0.log" Nov 27 18:13:57 crc kubenswrapper[4809]: I1127 18:13:57.954704 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-ll5l2_da239d06-b1c7-4b06-b830-5108030beed6/manager/0.log" Nov 27 18:13:58 crc kubenswrapper[4809]: I1127 18:13:58.023315 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-xrjzd_1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659/kube-rbac-proxy/0.log" Nov 27 18:13:58 crc kubenswrapper[4809]: I1127 18:13:58.042107 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-xrjzd_1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659/manager/0.log" Nov 27 18:13:58 crc kubenswrapper[4809]: I1127 18:13:58.176210 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm_c077fc08-1ee2-493e-b631-f01ccd0b7c6e/kube-rbac-proxy/0.log" Nov 27 18:13:58 crc kubenswrapper[4809]: I1127 18:13:58.219908 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm_c077fc08-1ee2-493e-b631-f01ccd0b7c6e/manager/0.log" Nov 27 18:13:58 crc kubenswrapper[4809]: I1127 18:13:58.546971 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-ckqts_45b7ef24-d97f-4dbe-8ad2-82f7f9983c68/registry-server/0.log" Nov 27 18:13:58 crc kubenswrapper[4809]: I1127 18:13:58.565401 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-b44dff85c-jvvsd_cceba3cb-7f9b-49d8-96e4-3e0e3927d106/operator/0.log" Nov 27 18:13:58 crc kubenswrapper[4809]: I1127 18:13:58.804577 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-c9pds_8ace4cdc-1253-4a0f-b667-83e997947f9b/kube-rbac-proxy/0.log" Nov 27 18:13:58 crc kubenswrapper[4809]: I1127 18:13:58.856321 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-c9pds_8ace4cdc-1253-4a0f-b667-83e997947f9b/manager/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.056094 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-8bq9l_3da94235-386c-4f05-b869-2e08cee40d5d/kube-rbac-proxy/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.128039 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-8bq9l_3da94235-386c-4f05-b869-2e08cee40d5d/manager/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.163517 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5b9vp_21cb4ed4-198d-4dd0-b962-311bf2bcc850/operator/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.301489 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-vvtmf_8f8e02f4-c34c-4b0a-b18f-8e088d48e12c/kube-rbac-proxy/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.392736 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-vvtmf_8f8e02f4-c34c-4b0a-b18f-8e088d48e12c/manager/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.404521 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6644d5b8df-l9kv7_18ae1240-d812-4f11-b515-6e7a94b8468e/manager/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.478418 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-ff79b6df5-lq72j_6016f85d-a0f4-45ee-a96f-95dcb2c57e3c/kube-rbac-proxy/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.608145 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-j6nv4_8798a938-aa20-40e2-be26-408026fb123e/kube-rbac-proxy/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.622118 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-ff79b6df5-lq72j_6016f85d-a0f4-45ee-a96f-95dcb2c57e3c/manager/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.642386 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-j6nv4_8798a938-aa20-40e2-be26-408026fb123e/manager/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.775709 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-d5hqw_5d49d329-6737-4bab-b243-64e618c0fa3b/kube-rbac-proxy/0.log" Nov 27 18:13:59 crc kubenswrapper[4809]: I1127 18:13:59.782354 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-d5hqw_5d49d329-6737-4bab-b243-64e618c0fa3b/manager/0.log" Nov 27 18:14:05 crc kubenswrapper[4809]: I1127 18:14:05.465143 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:14:05 crc kubenswrapper[4809]: E1127 18:14:05.465945 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:14:15 crc kubenswrapper[4809]: I1127 18:14:15.629859 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-tvl4x_87e2c04c-53e6-4524-90b8-b122e31b9280/control-plane-machine-set-operator/0.log" Nov 27 18:14:15 crc kubenswrapper[4809]: I1127 18:14:15.762096 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6rxcc_6d8b54b5-a20b-4461-bca8-c8ac2ec82048/kube-rbac-proxy/0.log" Nov 27 18:14:15 crc kubenswrapper[4809]: I1127 18:14:15.776981 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6rxcc_6d8b54b5-a20b-4461-bca8-c8ac2ec82048/machine-api-operator/0.log" Nov 27 18:14:17 crc kubenswrapper[4809]: I1127 18:14:17.457775 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:14:17 crc kubenswrapper[4809]: E1127 18:14:17.458311 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:14:29 crc kubenswrapper[4809]: I1127 18:14:29.409807 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7hzbp_df88093a-9522-45ca-913a-18a62128af63/cert-manager-controller/0.log" Nov 27 18:14:29 crc kubenswrapper[4809]: I1127 18:14:29.924437 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-tkpwp_79bf3153-df84-43fd-baa1-731fbc7f7873/cert-manager-webhook/0.log" Nov 27 18:14:29 crc kubenswrapper[4809]: I1127 18:14:29.942510 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rczjv_04d521a8-9cbf-4ca8-b7a8-035504c16c17/cert-manager-cainjector/0.log" Nov 27 18:14:32 crc kubenswrapper[4809]: I1127 18:14:32.458141 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:14:32 crc kubenswrapper[4809]: E1127 18:14:32.459242 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:14:42 crc kubenswrapper[4809]: I1127 18:14:42.409948 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-7lwv9_e0648dfc-fcc2-49c9-863b-0c40c33ea417/nmstate-console-plugin/0.log" Nov 27 18:14:42 crc kubenswrapper[4809]: I1127 18:14:42.560078 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-sccnl_acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b/nmstate-handler/0.log" Nov 27 18:14:42 crc kubenswrapper[4809]: I1127 18:14:42.592552 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-52jgm_637b5bb3-b183-43dd-bca4-0349ac11e7d5/kube-rbac-proxy/0.log" Nov 27 18:14:42 crc kubenswrapper[4809]: I1127 18:14:42.614166 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-52jgm_637b5bb3-b183-43dd-bca4-0349ac11e7d5/nmstate-metrics/0.log" Nov 27 18:14:42 crc kubenswrapper[4809]: I1127 18:14:42.783694 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-pz6jf_aca17653-5316-4c78-8062-9526abf28d97/nmstate-operator/0.log" Nov 27 18:14:42 crc kubenswrapper[4809]: I1127 18:14:42.836221 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-h86r5_796eb450-becc-4200-87e8-a9274d390901/nmstate-webhook/0.log" Nov 27 18:14:46 crc kubenswrapper[4809]: I1127 18:14:46.458360 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:14:46 crc kubenswrapper[4809]: E1127 18:14:46.459167 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:14:54 crc kubenswrapper[4809]: I1127 18:14:54.134511 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-fbb7c7d9f-bnjq7_2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61/kube-rbac-proxy/0.log" Nov 27 18:14:54 crc kubenswrapper[4809]: I1127 18:14:54.172260 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-fbb7c7d9f-bnjq7_2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61/manager/0.log" Nov 27 18:14:59 crc kubenswrapper[4809]: I1127 18:14:59.457588 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:14:59 crc kubenswrapper[4809]: E1127 18:14:59.459368 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.181348 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb"] Nov 27 18:15:00 crc kubenswrapper[4809]: E1127 18:15:00.181848 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerName="extract-utilities" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.181869 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerName="extract-utilities" Nov 27 18:15:00 crc kubenswrapper[4809]: E1127 18:15:00.181886 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerName="extract-content" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.181893 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerName="extract-content" Nov 27 18:15:00 crc kubenswrapper[4809]: E1127 18:15:00.181907 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerName="registry-server" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.181915 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerName="registry-server" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.182118 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ccebb66-4111-4960-ba0c-4510f5ce941b" containerName="registry-server" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.182889 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.184788 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.185065 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.193928 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb"] Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.264695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqcn7\" (UniqueName: \"kubernetes.io/projected/c946dce6-e7d2-464a-9749-f4bde497e436-kube-api-access-cqcn7\") pod \"collect-profiles-29404455-zdjtb\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.265124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c946dce6-e7d2-464a-9749-f4bde497e436-config-volume\") pod \"collect-profiles-29404455-zdjtb\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.265234 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c946dce6-e7d2-464a-9749-f4bde497e436-secret-volume\") pod \"collect-profiles-29404455-zdjtb\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.366825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c946dce6-e7d2-464a-9749-f4bde497e436-secret-volume\") pod \"collect-profiles-29404455-zdjtb\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.367038 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqcn7\" (UniqueName: \"kubernetes.io/projected/c946dce6-e7d2-464a-9749-f4bde497e436-kube-api-access-cqcn7\") pod \"collect-profiles-29404455-zdjtb\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.367082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c946dce6-e7d2-464a-9749-f4bde497e436-config-volume\") pod \"collect-profiles-29404455-zdjtb\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.368325 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c946dce6-e7d2-464a-9749-f4bde497e436-config-volume\") pod \"collect-profiles-29404455-zdjtb\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.374194 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c946dce6-e7d2-464a-9749-f4bde497e436-secret-volume\") pod \"collect-profiles-29404455-zdjtb\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.390642 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqcn7\" (UniqueName: \"kubernetes.io/projected/c946dce6-e7d2-464a-9749-f4bde497e436-kube-api-access-cqcn7\") pod \"collect-profiles-29404455-zdjtb\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.512319 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:00 crc kubenswrapper[4809]: I1127 18:15:00.968570 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb"] Nov 27 18:15:01 crc kubenswrapper[4809]: I1127 18:15:01.907261 4809 generic.go:334] "Generic (PLEG): container finished" podID="c946dce6-e7d2-464a-9749-f4bde497e436" containerID="6aacf06fa2caf0b231698f9bdd8aa90f06b01ba15ff3a9f703c0181e82f70f6d" exitCode=0 Nov 27 18:15:01 crc kubenswrapper[4809]: I1127 18:15:01.908074 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" event={"ID":"c946dce6-e7d2-464a-9749-f4bde497e436","Type":"ContainerDied","Data":"6aacf06fa2caf0b231698f9bdd8aa90f06b01ba15ff3a9f703c0181e82f70f6d"} Nov 27 18:15:01 crc kubenswrapper[4809]: I1127 18:15:01.908102 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" event={"ID":"c946dce6-e7d2-464a-9749-f4bde497e436","Type":"ContainerStarted","Data":"a5bfe1226eecc79dfed81240b32976de1fcaba7ee4a03cf28087d818b03197ea"} Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.330137 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.429134 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c946dce6-e7d2-464a-9749-f4bde497e436-config-volume\") pod \"c946dce6-e7d2-464a-9749-f4bde497e436\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.429201 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c946dce6-e7d2-464a-9749-f4bde497e436-secret-volume\") pod \"c946dce6-e7d2-464a-9749-f4bde497e436\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.429269 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqcn7\" (UniqueName: \"kubernetes.io/projected/c946dce6-e7d2-464a-9749-f4bde497e436-kube-api-access-cqcn7\") pod \"c946dce6-e7d2-464a-9749-f4bde497e436\" (UID: \"c946dce6-e7d2-464a-9749-f4bde497e436\") " Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.429806 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c946dce6-e7d2-464a-9749-f4bde497e436-config-volume" (OuterVolumeSpecName: "config-volume") pod "c946dce6-e7d2-464a-9749-f4bde497e436" (UID: "c946dce6-e7d2-464a-9749-f4bde497e436"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.434682 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c946dce6-e7d2-464a-9749-f4bde497e436-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c946dce6-e7d2-464a-9749-f4bde497e436" (UID: "c946dce6-e7d2-464a-9749-f4bde497e436"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.436013 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c946dce6-e7d2-464a-9749-f4bde497e436-kube-api-access-cqcn7" (OuterVolumeSpecName: "kube-api-access-cqcn7") pod "c946dce6-e7d2-464a-9749-f4bde497e436" (UID: "c946dce6-e7d2-464a-9749-f4bde497e436"). InnerVolumeSpecName "kube-api-access-cqcn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.531692 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c946dce6-e7d2-464a-9749-f4bde497e436-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.531725 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c946dce6-e7d2-464a-9749-f4bde497e436-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.531733 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqcn7\" (UniqueName: \"kubernetes.io/projected/c946dce6-e7d2-464a-9749-f4bde497e436-kube-api-access-cqcn7\") on node \"crc\" DevicePath \"\"" Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.926733 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" event={"ID":"c946dce6-e7d2-464a-9749-f4bde497e436","Type":"ContainerDied","Data":"a5bfe1226eecc79dfed81240b32976de1fcaba7ee4a03cf28087d818b03197ea"} Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.927083 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5bfe1226eecc79dfed81240b32976de1fcaba7ee4a03cf28087d818b03197ea" Nov 27 18:15:03 crc kubenswrapper[4809]: I1127 18:15:03.926800 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404455-zdjtb" Nov 27 18:15:04 crc kubenswrapper[4809]: I1127 18:15:04.415245 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp"] Nov 27 18:15:04 crc kubenswrapper[4809]: I1127 18:15:04.427251 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404410-s7mpp"] Nov 27 18:15:05 crc kubenswrapper[4809]: I1127 18:15:05.472702 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb1409d8-a556-486c-8a94-8e5caf14f196" path="/var/lib/kubelet/pods/bb1409d8-a556-486c-8a94-8e5caf14f196/volumes" Nov 27 18:15:07 crc kubenswrapper[4809]: I1127 18:15:07.375639 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-8zwm2_c0697c42-98e3-49fd-b923-093806ec2876/cluster-logging-operator/0.log" Nov 27 18:15:07 crc kubenswrapper[4809]: I1127 18:15:07.601625 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-ndnhl_1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1/collector/0.log" Nov 27 18:15:07 crc kubenswrapper[4809]: I1127 18:15:07.610282 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_ed55809b-892d-4ba2-b037-5eb25f1add8d/loki-compactor/0.log" Nov 27 18:15:07 crc kubenswrapper[4809]: I1127 18:15:07.793261 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-wmb6l_7d6e72ed-2edc-4744-8b78-adb9d95a41ee/loki-distributor/0.log" Nov 27 18:15:07 crc kubenswrapper[4809]: I1127 18:15:07.837461 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-86596f6498-4d4pc_4239a407-a58d-46ac-a13b-0f744c3fbf0f/opa/0.log" Nov 27 18:15:07 crc kubenswrapper[4809]: I1127 18:15:07.838606 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-86596f6498-4d4pc_4239a407-a58d-46ac-a13b-0f744c3fbf0f/gateway/0.log" Nov 27 18:15:08 crc kubenswrapper[4809]: I1127 18:15:08.380858 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-86596f6498-t5b62_42bade25-cca5-4a1b-9c2f-67c1ffb97c68/gateway/0.log" Nov 27 18:15:08 crc kubenswrapper[4809]: I1127 18:15:08.381288 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-86596f6498-t5b62_42bade25-cca5-4a1b-9c2f-67c1ffb97c68/opa/0.log" Nov 27 18:15:08 crc kubenswrapper[4809]: I1127 18:15:08.564851 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_cccadb3e-654a-430f-9c69-c99a6fde2279/loki-index-gateway/0.log" Nov 27 18:15:08 crc kubenswrapper[4809]: I1127 18:15:08.576530 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_87ecc35f-bd3c-43be-9841-0129299f73f9/loki-ingester/0.log" Nov 27 18:15:08 crc kubenswrapper[4809]: I1127 18:15:08.717075 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-gr498_ce25b13d-dcd3-420d-848a-afd09c1850f2/loki-querier/0.log" Nov 27 18:15:08 crc kubenswrapper[4809]: I1127 18:15:08.760427 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-flnsq_290cdb65-2424-4494-94cd-a69e5bbc126c/loki-query-frontend/0.log" Nov 27 18:15:10 crc kubenswrapper[4809]: I1127 18:15:10.458439 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:15:10 crc kubenswrapper[4809]: E1127 18:15:10.458691 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.381746 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rbkql_976b64ef-1aff-4780-8353-dde2792907d4/kube-rbac-proxy/0.log" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.507973 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rbkql_976b64ef-1aff-4780-8353-dde2792907d4/controller/0.log" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.598009 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-frr-files/0.log" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.733165 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-frr-files/0.log" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.788623 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-metrics/0.log" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.790781 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-reloader/0.log" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.822903 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-reloader/0.log" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.950600 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-metrics/0.log" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.974485 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-reloader/0.log" Nov 27 18:15:20 crc kubenswrapper[4809]: I1127 18:15:20.990014 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-frr-files/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.004337 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-metrics/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.165589 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-reloader/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.183161 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-frr-files/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.183221 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-metrics/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.187585 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/controller/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.360915 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/frr-metrics/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.363336 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/kube-rbac-proxy-frr/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.388683 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/kube-rbac-proxy/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.549427 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/reloader/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.584353 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-p48r2_42ffbf32-2946-4087-8474-c0c0ecead8eb/frr-k8s-webhook-server/0.log" Nov 27 18:15:21 crc kubenswrapper[4809]: I1127 18:15:21.835840 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-77cc889676-8xfh7_eca3efdb-8eaf-4bd0-a2a4-471180514ee5/manager/0.log" Nov 27 18:15:22 crc kubenswrapper[4809]: I1127 18:15:22.016503 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b59cf8cdf-wz5qf_779ed7ea-3841-4937-978f-046fe79f4a9c/webhook-server/0.log" Nov 27 18:15:22 crc kubenswrapper[4809]: I1127 18:15:22.108020 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-gwn22_283ae992-ff54-437d-aba9-3dc45f78d876/kube-rbac-proxy/0.log" Nov 27 18:15:22 crc kubenswrapper[4809]: I1127 18:15:22.674646 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-gwn22_283ae992-ff54-437d-aba9-3dc45f78d876/speaker/0.log" Nov 27 18:15:22 crc kubenswrapper[4809]: I1127 18:15:22.702002 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/frr/0.log" Nov 27 18:15:23 crc kubenswrapper[4809]: I1127 18:15:23.458801 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:15:23 crc kubenswrapper[4809]: E1127 18:15:23.459259 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:15:35 crc kubenswrapper[4809]: I1127 18:15:35.547569 4809 scope.go:117] "RemoveContainer" containerID="c779647fad2e5a6faeed63d86d563effd014d8ffe393623c81d00fe6a316d4c1" Nov 27 18:15:35 crc kubenswrapper[4809]: I1127 18:15:35.724650 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/util/0.log" Nov 27 18:15:35 crc kubenswrapper[4809]: I1127 18:15:35.909762 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/util/0.log" Nov 27 18:15:35 crc kubenswrapper[4809]: I1127 18:15:35.943394 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/pull/0.log" Nov 27 18:15:35 crc kubenswrapper[4809]: I1127 18:15:35.943536 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/pull/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.091478 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/util/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.118188 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/pull/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.122548 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/extract/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.246662 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/util/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.412528 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/util/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.419783 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/pull/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.424018 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/pull/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.457912 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:15:36 crc kubenswrapper[4809]: E1127 18:15:36.458160 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.946555 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/util/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.983423 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/extract/0.log" Nov 27 18:15:36 crc kubenswrapper[4809]: I1127 18:15:36.989923 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/pull/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.131826 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-utilities/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.263799 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-utilities/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.296021 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-content/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.298830 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-content/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.476584 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-content/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.490440 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-utilities/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.678965 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-utilities/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.919997 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-content/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.922105 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-utilities/0.log" Nov 27 18:15:37 crc kubenswrapper[4809]: I1127 18:15:37.944283 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-content/0.log" Nov 27 18:15:38 crc kubenswrapper[4809]: I1127 18:15:38.053260 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/registry-server/0.log" Nov 27 18:15:38 crc kubenswrapper[4809]: I1127 18:15:38.175533 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-content/0.log" Nov 27 18:15:38 crc kubenswrapper[4809]: I1127 18:15:38.178936 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-utilities/0.log" Nov 27 18:15:38 crc kubenswrapper[4809]: I1127 18:15:38.434883 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/registry-server/0.log" Nov 27 18:15:38 crc kubenswrapper[4809]: I1127 18:15:38.858350 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4x2bl_b3ffdbfe-1e8b-415f-9d97-47a17d84070c/marketplace-operator/0.log" Nov 27 18:15:38 crc kubenswrapper[4809]: I1127 18:15:38.864705 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-utilities/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.071289 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-utilities/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.071655 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-content/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.103075 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-content/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.236971 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-utilities/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.297787 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-content/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.402553 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-utilities/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.446585 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/registry-server/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.552054 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-utilities/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.572552 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-content/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.595732 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-content/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.750181 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-utilities/0.log" Nov 27 18:15:39 crc kubenswrapper[4809]: I1127 18:15:39.751507 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-content/0.log" Nov 27 18:15:40 crc kubenswrapper[4809]: I1127 18:15:40.235039 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/registry-server/0.log" Nov 27 18:15:48 crc kubenswrapper[4809]: I1127 18:15:48.457805 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:15:48 crc kubenswrapper[4809]: E1127 18:15:48.458443 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:16:01 crc kubenswrapper[4809]: I1127 18:16:01.457871 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:16:01 crc kubenswrapper[4809]: E1127 18:16:01.460008 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:16:02 crc kubenswrapper[4809]: I1127 18:16:02.519523 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-fbb7c7d9f-bnjq7_2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61/kube-rbac-proxy/0.log" Nov 27 18:16:02 crc kubenswrapper[4809]: I1127 18:16:02.617985 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-fbb7c7d9f-bnjq7_2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61/manager/0.log" Nov 27 18:16:16 crc kubenswrapper[4809]: I1127 18:16:16.458460 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:16:16 crc kubenswrapper[4809]: E1127 18:16:16.459146 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:16:29 crc kubenswrapper[4809]: I1127 18:16:29.464308 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:16:29 crc kubenswrapper[4809]: E1127 18:16:29.464996 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:16:43 crc kubenswrapper[4809]: I1127 18:16:43.458438 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:16:43 crc kubenswrapper[4809]: E1127 18:16:43.459326 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:16:58 crc kubenswrapper[4809]: I1127 18:16:58.458989 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:16:58 crc kubenswrapper[4809]: E1127 18:16:58.460020 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:17:12 crc kubenswrapper[4809]: I1127 18:17:12.458634 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:17:12 crc kubenswrapper[4809]: E1127 18:17:12.459692 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:17:24 crc kubenswrapper[4809]: I1127 18:17:24.458130 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:17:24 crc kubenswrapper[4809]: E1127 18:17:24.459170 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:17:36 crc kubenswrapper[4809]: I1127 18:17:36.459020 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:17:36 crc kubenswrapper[4809]: E1127 18:17:36.460001 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:17:37 crc kubenswrapper[4809]: I1127 18:17:37.581803 4809 generic.go:334] "Generic (PLEG): container finished" podID="fa97813a-2b2c-43a3-9084-06f4ae660067" containerID="90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60" exitCode=0 Nov 27 18:17:37 crc kubenswrapper[4809]: I1127 18:17:37.581891 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9khf6/must-gather-p5rhm" event={"ID":"fa97813a-2b2c-43a3-9084-06f4ae660067","Type":"ContainerDied","Data":"90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60"} Nov 27 18:17:37 crc kubenswrapper[4809]: I1127 18:17:37.582657 4809 scope.go:117] "RemoveContainer" containerID="90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60" Nov 27 18:17:37 crc kubenswrapper[4809]: I1127 18:17:37.815316 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9khf6_must-gather-p5rhm_fa97813a-2b2c-43a3-9084-06f4ae660067/gather/0.log" Nov 27 18:17:46 crc kubenswrapper[4809]: I1127 18:17:46.968156 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9khf6/must-gather-p5rhm"] Nov 27 18:17:46 crc kubenswrapper[4809]: I1127 18:17:46.969942 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-9khf6/must-gather-p5rhm" podUID="fa97813a-2b2c-43a3-9084-06f4ae660067" containerName="copy" containerID="cri-o://dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1" gracePeriod=2 Nov 27 18:17:46 crc kubenswrapper[4809]: I1127 18:17:46.979985 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9khf6/must-gather-p5rhm"] Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.435686 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9khf6_must-gather-p5rhm_fa97813a-2b2c-43a3-9084-06f4ae660067/copy/0.log" Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.438451 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.597401 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa97813a-2b2c-43a3-9084-06f4ae660067-must-gather-output\") pod \"fa97813a-2b2c-43a3-9084-06f4ae660067\" (UID: \"fa97813a-2b2c-43a3-9084-06f4ae660067\") " Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.597717 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4dqz\" (UniqueName: \"kubernetes.io/projected/fa97813a-2b2c-43a3-9084-06f4ae660067-kube-api-access-t4dqz\") pod \"fa97813a-2b2c-43a3-9084-06f4ae660067\" (UID: \"fa97813a-2b2c-43a3-9084-06f4ae660067\") " Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.603208 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa97813a-2b2c-43a3-9084-06f4ae660067-kube-api-access-t4dqz" (OuterVolumeSpecName: "kube-api-access-t4dqz") pod "fa97813a-2b2c-43a3-9084-06f4ae660067" (UID: "fa97813a-2b2c-43a3-9084-06f4ae660067"). InnerVolumeSpecName "kube-api-access-t4dqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.689031 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9khf6_must-gather-p5rhm_fa97813a-2b2c-43a3-9084-06f4ae660067/copy/0.log" Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.689480 4809 generic.go:334] "Generic (PLEG): container finished" podID="fa97813a-2b2c-43a3-9084-06f4ae660067" containerID="dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1" exitCode=143 Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.689532 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9khf6/must-gather-p5rhm" Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.689538 4809 scope.go:117] "RemoveContainer" containerID="dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1" Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.701384 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4dqz\" (UniqueName: \"kubernetes.io/projected/fa97813a-2b2c-43a3-9084-06f4ae660067-kube-api-access-t4dqz\") on node \"crc\" DevicePath \"\"" Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.716046 4809 scope.go:117] "RemoveContainer" containerID="90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60" Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.742826 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa97813a-2b2c-43a3-9084-06f4ae660067-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fa97813a-2b2c-43a3-9084-06f4ae660067" (UID: "fa97813a-2b2c-43a3-9084-06f4ae660067"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:17:47 crc kubenswrapper[4809]: I1127 18:17:47.803035 4809 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa97813a-2b2c-43a3-9084-06f4ae660067-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 27 18:17:48 crc kubenswrapper[4809]: I1127 18:17:48.381457 4809 scope.go:117] "RemoveContainer" containerID="dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1" Nov 27 18:17:48 crc kubenswrapper[4809]: E1127 18:17:48.382347 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1\": container with ID starting with dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1 not found: ID does not exist" containerID="dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1" Nov 27 18:17:48 crc kubenswrapper[4809]: I1127 18:17:48.382387 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1"} err="failed to get container status \"dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1\": rpc error: code = NotFound desc = could not find container \"dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1\": container with ID starting with dcad6143eb3c280572184202d012656ab1499307da68174abd0750049f9907d1 not found: ID does not exist" Nov 27 18:17:48 crc kubenswrapper[4809]: I1127 18:17:48.382414 4809 scope.go:117] "RemoveContainer" containerID="90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60" Nov 27 18:17:48 crc kubenswrapper[4809]: E1127 18:17:48.382811 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60\": container with ID starting with 90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60 not found: ID does not exist" containerID="90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60" Nov 27 18:17:48 crc kubenswrapper[4809]: I1127 18:17:48.382848 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60"} err="failed to get container status \"90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60\": rpc error: code = NotFound desc = could not find container \"90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60\": container with ID starting with 90710c89ca0d27aa80a9db17e6e9c922dc9c860e0355c3642dc8ac21203cce60 not found: ID does not exist" Nov 27 18:17:49 crc kubenswrapper[4809]: I1127 18:17:49.469845 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa97813a-2b2c-43a3-9084-06f4ae660067" path="/var/lib/kubelet/pods/fa97813a-2b2c-43a3-9084-06f4ae660067/volumes" Nov 27 18:17:50 crc kubenswrapper[4809]: I1127 18:17:50.458848 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:17:50 crc kubenswrapper[4809]: E1127 18:17:50.459553 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:18:01 crc kubenswrapper[4809]: I1127 18:18:01.457673 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:18:01 crc kubenswrapper[4809]: E1127 18:18:01.458492 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:18:14 crc kubenswrapper[4809]: I1127 18:18:14.457611 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:18:14 crc kubenswrapper[4809]: E1127 18:18:14.459023 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.330598 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hngdc"] Nov 27 18:18:18 crc kubenswrapper[4809]: E1127 18:18:18.331403 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa97813a-2b2c-43a3-9084-06f4ae660067" containerName="gather" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.331422 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa97813a-2b2c-43a3-9084-06f4ae660067" containerName="gather" Nov 27 18:18:18 crc kubenswrapper[4809]: E1127 18:18:18.331461 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c946dce6-e7d2-464a-9749-f4bde497e436" containerName="collect-profiles" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.331468 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c946dce6-e7d2-464a-9749-f4bde497e436" containerName="collect-profiles" Nov 27 18:18:18 crc kubenswrapper[4809]: E1127 18:18:18.331504 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa97813a-2b2c-43a3-9084-06f4ae660067" containerName="copy" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.331510 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa97813a-2b2c-43a3-9084-06f4ae660067" containerName="copy" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.331705 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa97813a-2b2c-43a3-9084-06f4ae660067" containerName="copy" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.331721 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa97813a-2b2c-43a3-9084-06f4ae660067" containerName="gather" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.331733 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c946dce6-e7d2-464a-9749-f4bde497e436" containerName="collect-profiles" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.333254 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.340411 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hngdc"] Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.501886 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-catalog-content\") pod \"redhat-operators-hngdc\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.502233 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96pd8\" (UniqueName: \"kubernetes.io/projected/58b7e21a-bf60-472e-bd23-6033b4e880c7-kube-api-access-96pd8\") pod \"redhat-operators-hngdc\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.502301 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-utilities\") pod \"redhat-operators-hngdc\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.604200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-catalog-content\") pod \"redhat-operators-hngdc\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.604564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96pd8\" (UniqueName: \"kubernetes.io/projected/58b7e21a-bf60-472e-bd23-6033b4e880c7-kube-api-access-96pd8\") pod \"redhat-operators-hngdc\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.604689 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-catalog-content\") pod \"redhat-operators-hngdc\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.604867 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-utilities\") pod \"redhat-operators-hngdc\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.605359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-utilities\") pod \"redhat-operators-hngdc\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.710921 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96pd8\" (UniqueName: \"kubernetes.io/projected/58b7e21a-bf60-472e-bd23-6033b4e880c7-kube-api-access-96pd8\") pod \"redhat-operators-hngdc\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:18 crc kubenswrapper[4809]: I1127 18:18:18.960353 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:19 crc kubenswrapper[4809]: I1127 18:18:19.676903 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hngdc"] Nov 27 18:18:19 crc kubenswrapper[4809]: I1127 18:18:19.957234 4809 generic.go:334] "Generic (PLEG): container finished" podID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerID="dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385" exitCode=0 Nov 27 18:18:19 crc kubenswrapper[4809]: I1127 18:18:19.957514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hngdc" event={"ID":"58b7e21a-bf60-472e-bd23-6033b4e880c7","Type":"ContainerDied","Data":"dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385"} Nov 27 18:18:19 crc kubenswrapper[4809]: I1127 18:18:19.957543 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hngdc" event={"ID":"58b7e21a-bf60-472e-bd23-6033b4e880c7","Type":"ContainerStarted","Data":"214adc56c148396ca8848f73d89030d8e1465521b6660261ae368692a1ef8c07"} Nov 27 18:18:19 crc kubenswrapper[4809]: I1127 18:18:19.959167 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 18:18:20 crc kubenswrapper[4809]: I1127 18:18:20.968362 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hngdc" event={"ID":"58b7e21a-bf60-472e-bd23-6033b4e880c7","Type":"ContainerStarted","Data":"7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc"} Nov 27 18:18:25 crc kubenswrapper[4809]: I1127 18:18:25.006159 4809 generic.go:334] "Generic (PLEG): container finished" podID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerID="7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc" exitCode=0 Nov 27 18:18:25 crc kubenswrapper[4809]: I1127 18:18:25.006241 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hngdc" event={"ID":"58b7e21a-bf60-472e-bd23-6033b4e880c7","Type":"ContainerDied","Data":"7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc"} Nov 27 18:18:27 crc kubenswrapper[4809]: I1127 18:18:27.027156 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hngdc" event={"ID":"58b7e21a-bf60-472e-bd23-6033b4e880c7","Type":"ContainerStarted","Data":"d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a"} Nov 27 18:18:27 crc kubenswrapper[4809]: I1127 18:18:27.052222 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hngdc" podStartSLOduration=3.236551113 podStartE2EDuration="9.052200889s" podCreationTimestamp="2025-11-27 18:18:18 +0000 UTC" firstStartedPulling="2025-11-27 18:18:19.958983005 +0000 UTC m=+4135.231440347" lastFinishedPulling="2025-11-27 18:18:25.774632771 +0000 UTC m=+4141.047090123" observedRunningTime="2025-11-27 18:18:27.042201286 +0000 UTC m=+4142.314658648" watchObservedRunningTime="2025-11-27 18:18:27.052200889 +0000 UTC m=+4142.324658241" Nov 27 18:18:28 crc kubenswrapper[4809]: I1127 18:18:28.458609 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:18:28 crc kubenswrapper[4809]: I1127 18:18:28.964861 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:28 crc kubenswrapper[4809]: I1127 18:18:28.965219 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:29 crc kubenswrapper[4809]: I1127 18:18:29.047166 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"7b000adc98c38ad552afdfcf0fbc8fde8da0963aa23c44e43a46c14ec901bccf"} Nov 27 18:18:30 crc kubenswrapper[4809]: I1127 18:18:30.029467 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hngdc" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerName="registry-server" probeResult="failure" output=< Nov 27 18:18:30 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 27 18:18:30 crc kubenswrapper[4809]: > Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.032020 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wbvfh"] Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.034621 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.051015 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wbvfh"] Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.077447 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.108394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skb9f\" (UniqueName: \"kubernetes.io/projected/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-kube-api-access-skb9f\") pod \"certified-operators-wbvfh\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.108534 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-utilities\") pod \"certified-operators-wbvfh\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.108564 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-catalog-content\") pod \"certified-operators-wbvfh\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.134241 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.210762 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skb9f\" (UniqueName: \"kubernetes.io/projected/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-kube-api-access-skb9f\") pod \"certified-operators-wbvfh\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.210889 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-utilities\") pod \"certified-operators-wbvfh\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.210912 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-catalog-content\") pod \"certified-operators-wbvfh\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.211548 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-catalog-content\") pod \"certified-operators-wbvfh\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.211544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-utilities\") pod \"certified-operators-wbvfh\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.234467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skb9f\" (UniqueName: \"kubernetes.io/projected/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-kube-api-access-skb9f\") pod \"certified-operators-wbvfh\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.355707 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:39 crc kubenswrapper[4809]: I1127 18:18:39.837708 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wbvfh"] Nov 27 18:18:40 crc kubenswrapper[4809]: I1127 18:18:40.146091 4809 generic.go:334] "Generic (PLEG): container finished" podID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerID="a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42" exitCode=0 Nov 27 18:18:40 crc kubenswrapper[4809]: I1127 18:18:40.146168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbvfh" event={"ID":"6e2ac114-e33e-4ec3-bda2-59b5fdc59076","Type":"ContainerDied","Data":"a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42"} Nov 27 18:18:40 crc kubenswrapper[4809]: I1127 18:18:40.146383 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbvfh" event={"ID":"6e2ac114-e33e-4ec3-bda2-59b5fdc59076","Type":"ContainerStarted","Data":"3c7ed67329b5ac271de25ad8940c582cc1dc5d23d5ceceb66487c6cd988dd5b2"} Nov 27 18:18:41 crc kubenswrapper[4809]: I1127 18:18:41.155882 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbvfh" event={"ID":"6e2ac114-e33e-4ec3-bda2-59b5fdc59076","Type":"ContainerStarted","Data":"1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068"} Nov 27 18:18:41 crc kubenswrapper[4809]: I1127 18:18:41.396811 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hngdc"] Nov 27 18:18:41 crc kubenswrapper[4809]: I1127 18:18:41.397139 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hngdc" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerName="registry-server" containerID="cri-o://d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a" gracePeriod=2 Nov 27 18:18:41 crc kubenswrapper[4809]: I1127 18:18:41.868397 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:41 crc kubenswrapper[4809]: I1127 18:18:41.965014 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-utilities\") pod \"58b7e21a-bf60-472e-bd23-6033b4e880c7\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " Nov 27 18:18:41 crc kubenswrapper[4809]: I1127 18:18:41.965393 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-catalog-content\") pod \"58b7e21a-bf60-472e-bd23-6033b4e880c7\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " Nov 27 18:18:41 crc kubenswrapper[4809]: I1127 18:18:41.965421 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96pd8\" (UniqueName: \"kubernetes.io/projected/58b7e21a-bf60-472e-bd23-6033b4e880c7-kube-api-access-96pd8\") pod \"58b7e21a-bf60-472e-bd23-6033b4e880c7\" (UID: \"58b7e21a-bf60-472e-bd23-6033b4e880c7\") " Nov 27 18:18:41 crc kubenswrapper[4809]: I1127 18:18:41.966039 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-utilities" (OuterVolumeSpecName: "utilities") pod "58b7e21a-bf60-472e-bd23-6033b4e880c7" (UID: "58b7e21a-bf60-472e-bd23-6033b4e880c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:18:41 crc kubenswrapper[4809]: I1127 18:18:41.971922 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58b7e21a-bf60-472e-bd23-6033b4e880c7-kube-api-access-96pd8" (OuterVolumeSpecName: "kube-api-access-96pd8") pod "58b7e21a-bf60-472e-bd23-6033b4e880c7" (UID: "58b7e21a-bf60-472e-bd23-6033b4e880c7"). InnerVolumeSpecName "kube-api-access-96pd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.068227 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.068278 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96pd8\" (UniqueName: \"kubernetes.io/projected/58b7e21a-bf60-472e-bd23-6033b4e880c7-kube-api-access-96pd8\") on node \"crc\" DevicePath \"\"" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.077624 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58b7e21a-bf60-472e-bd23-6033b4e880c7" (UID: "58b7e21a-bf60-472e-bd23-6033b4e880c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.166575 4809 generic.go:334] "Generic (PLEG): container finished" podID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerID="d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a" exitCode=0 Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.166615 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hngdc" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.166625 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hngdc" event={"ID":"58b7e21a-bf60-472e-bd23-6033b4e880c7","Type":"ContainerDied","Data":"d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a"} Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.166941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hngdc" event={"ID":"58b7e21a-bf60-472e-bd23-6033b4e880c7","Type":"ContainerDied","Data":"214adc56c148396ca8848f73d89030d8e1465521b6660261ae368692a1ef8c07"} Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.166980 4809 scope.go:117] "RemoveContainer" containerID="d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.168960 4809 generic.go:334] "Generic (PLEG): container finished" podID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerID="1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068" exitCode=0 Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.169113 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbvfh" event={"ID":"6e2ac114-e33e-4ec3-bda2-59b5fdc59076","Type":"ContainerDied","Data":"1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068"} Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.170780 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b7e21a-bf60-472e-bd23-6033b4e880c7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.192327 4809 scope.go:117] "RemoveContainer" containerID="7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.207698 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hngdc"] Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.215577 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hngdc"] Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.217604 4809 scope.go:117] "RemoveContainer" containerID="dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.266171 4809 scope.go:117] "RemoveContainer" containerID="d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a" Nov 27 18:18:42 crc kubenswrapper[4809]: E1127 18:18:42.266699 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a\": container with ID starting with d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a not found: ID does not exist" containerID="d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.266771 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a"} err="failed to get container status \"d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a\": rpc error: code = NotFound desc = could not find container \"d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a\": container with ID starting with d7c0ea0c100799460df8b4fb44becaac637e8196cb9770cfbf5dc91910e4e21a not found: ID does not exist" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.266804 4809 scope.go:117] "RemoveContainer" containerID="7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc" Nov 27 18:18:42 crc kubenswrapper[4809]: E1127 18:18:42.267248 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc\": container with ID starting with 7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc not found: ID does not exist" containerID="7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.267280 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc"} err="failed to get container status \"7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc\": rpc error: code = NotFound desc = could not find container \"7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc\": container with ID starting with 7a42be7e9dbfa307b8d49a08ceaedbcd5348d5658e4c9e34e5dedb39b4a189cc not found: ID does not exist" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.267302 4809 scope.go:117] "RemoveContainer" containerID="dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385" Nov 27 18:18:42 crc kubenswrapper[4809]: E1127 18:18:42.267694 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385\": container with ID starting with dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385 not found: ID does not exist" containerID="dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385" Nov 27 18:18:42 crc kubenswrapper[4809]: I1127 18:18:42.267727 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385"} err="failed to get container status \"dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385\": rpc error: code = NotFound desc = could not find container \"dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385\": container with ID starting with dd57a72782968389aaa763f5b10d7d2b002a19449bb992a4ed283274e37ad385 not found: ID does not exist" Nov 27 18:18:43 crc kubenswrapper[4809]: I1127 18:18:43.180615 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbvfh" event={"ID":"6e2ac114-e33e-4ec3-bda2-59b5fdc59076","Type":"ContainerStarted","Data":"8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb"} Nov 27 18:18:43 crc kubenswrapper[4809]: I1127 18:18:43.222920 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wbvfh" podStartSLOduration=1.459632204 podStartE2EDuration="4.222901462s" podCreationTimestamp="2025-11-27 18:18:39 +0000 UTC" firstStartedPulling="2025-11-27 18:18:40.147977998 +0000 UTC m=+4155.420435350" lastFinishedPulling="2025-11-27 18:18:42.911247256 +0000 UTC m=+4158.183704608" observedRunningTime="2025-11-27 18:18:43.197099358 +0000 UTC m=+4158.469556710" watchObservedRunningTime="2025-11-27 18:18:43.222901462 +0000 UTC m=+4158.495358844" Nov 27 18:18:43 crc kubenswrapper[4809]: I1127 18:18:43.470167 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" path="/var/lib/kubelet/pods/58b7e21a-bf60-472e-bd23-6033b4e880c7/volumes" Nov 27 18:18:49 crc kubenswrapper[4809]: I1127 18:18:49.356064 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:49 crc kubenswrapper[4809]: I1127 18:18:49.356507 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:49 crc kubenswrapper[4809]: I1127 18:18:49.408156 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:50 crc kubenswrapper[4809]: I1127 18:18:50.292036 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:50 crc kubenswrapper[4809]: I1127 18:18:50.344959 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wbvfh"] Nov 27 18:18:52 crc kubenswrapper[4809]: I1127 18:18:52.261526 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wbvfh" podUID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerName="registry-server" containerID="cri-o://8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb" gracePeriod=2 Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.194067 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.278540 4809 generic.go:334] "Generic (PLEG): container finished" podID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerID="8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb" exitCode=0 Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.278591 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbvfh" event={"ID":"6e2ac114-e33e-4ec3-bda2-59b5fdc59076","Type":"ContainerDied","Data":"8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb"} Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.278625 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbvfh" event={"ID":"6e2ac114-e33e-4ec3-bda2-59b5fdc59076","Type":"ContainerDied","Data":"3c7ed67329b5ac271de25ad8940c582cc1dc5d23d5ceceb66487c6cd988dd5b2"} Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.278649 4809 scope.go:117] "RemoveContainer" containerID="8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.278865 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbvfh" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.292370 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skb9f\" (UniqueName: \"kubernetes.io/projected/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-kube-api-access-skb9f\") pod \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.292519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-utilities\") pod \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.292619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-catalog-content\") pod \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\" (UID: \"6e2ac114-e33e-4ec3-bda2-59b5fdc59076\") " Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.301127 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-utilities" (OuterVolumeSpecName: "utilities") pod "6e2ac114-e33e-4ec3-bda2-59b5fdc59076" (UID: "6e2ac114-e33e-4ec3-bda2-59b5fdc59076"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.303450 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-kube-api-access-skb9f" (OuterVolumeSpecName: "kube-api-access-skb9f") pod "6e2ac114-e33e-4ec3-bda2-59b5fdc59076" (UID: "6e2ac114-e33e-4ec3-bda2-59b5fdc59076"). InnerVolumeSpecName "kube-api-access-skb9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.350920 4809 scope.go:117] "RemoveContainer" containerID="1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.395391 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skb9f\" (UniqueName: \"kubernetes.io/projected/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-kube-api-access-skb9f\") on node \"crc\" DevicePath \"\"" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.395431 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.402416 4809 scope.go:117] "RemoveContainer" containerID="a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.405237 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e2ac114-e33e-4ec3-bda2-59b5fdc59076" (UID: "6e2ac114-e33e-4ec3-bda2-59b5fdc59076"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.446505 4809 scope.go:117] "RemoveContainer" containerID="8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb" Nov 27 18:18:53 crc kubenswrapper[4809]: E1127 18:18:53.447170 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb\": container with ID starting with 8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb not found: ID does not exist" containerID="8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.447270 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb"} err="failed to get container status \"8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb\": rpc error: code = NotFound desc = could not find container \"8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb\": container with ID starting with 8d88b4d3f346972e495d5ad4b2741ecf626c697e5ff5931c32c7bb44235232cb not found: ID does not exist" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.447351 4809 scope.go:117] "RemoveContainer" containerID="1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068" Nov 27 18:18:53 crc kubenswrapper[4809]: E1127 18:18:53.447719 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068\": container with ID starting with 1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068 not found: ID does not exist" containerID="1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.447807 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068"} err="failed to get container status \"1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068\": rpc error: code = NotFound desc = could not find container \"1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068\": container with ID starting with 1d57933a9e4af7b83b57e896d86e37ab920524c605b20ec8bc9f7a6dc3627068 not found: ID does not exist" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.447839 4809 scope.go:117] "RemoveContainer" containerID="a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42" Nov 27 18:18:53 crc kubenswrapper[4809]: E1127 18:18:53.448186 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42\": container with ID starting with a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42 not found: ID does not exist" containerID="a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.448269 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42"} err="failed to get container status \"a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42\": rpc error: code = NotFound desc = could not find container \"a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42\": container with ID starting with a79856de4c7abc1b206c43a8d5fe0a21cc021a8f2b1e84532fb2e4a0cc03bc42 not found: ID does not exist" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.497624 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e2ac114-e33e-4ec3-bda2-59b5fdc59076-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.611028 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wbvfh"] Nov 27 18:18:53 crc kubenswrapper[4809]: I1127 18:18:53.620200 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wbvfh"] Nov 27 18:18:53 crc kubenswrapper[4809]: E1127 18:18:53.681644 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e2ac114_e33e_4ec3_bda2_59b5fdc59076.slice/crio-3c7ed67329b5ac271de25ad8940c582cc1dc5d23d5ceceb66487c6cd988dd5b2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e2ac114_e33e_4ec3_bda2_59b5fdc59076.slice\": RecentStats: unable to find data in memory cache]" Nov 27 18:18:55 crc kubenswrapper[4809]: I1127 18:18:55.469848 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" path="/var/lib/kubelet/pods/6e2ac114-e33e-4ec3-bda2-59b5fdc59076/volumes" Nov 27 18:19:35 crc kubenswrapper[4809]: I1127 18:19:35.684164 4809 scope.go:117] "RemoveContainer" containerID="2099aad695c2e550e3c372c3f8d3c82f3d199edee6002343567586ee536c352e" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.165208 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2sl"] Nov 27 18:19:52 crc kubenswrapper[4809]: E1127 18:19:52.166212 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerName="registry-server" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.166231 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerName="registry-server" Nov 27 18:19:52 crc kubenswrapper[4809]: E1127 18:19:52.166247 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerName="extract-content" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.166255 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerName="extract-content" Nov 27 18:19:52 crc kubenswrapper[4809]: E1127 18:19:52.166273 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerName="extract-utilities" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.166280 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerName="extract-utilities" Nov 27 18:19:52 crc kubenswrapper[4809]: E1127 18:19:52.166297 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerName="registry-server" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.166303 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerName="registry-server" Nov 27 18:19:52 crc kubenswrapper[4809]: E1127 18:19:52.166318 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerName="extract-utilities" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.166324 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerName="extract-utilities" Nov 27 18:19:52 crc kubenswrapper[4809]: E1127 18:19:52.166335 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerName="extract-content" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.166342 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerName="extract-content" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.166529 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e2ac114-e33e-4ec3-bda2-59b5fdc59076" containerName="registry-server" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.166544 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58b7e21a-bf60-472e-bd23-6033b4e880c7" containerName="registry-server" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.169787 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.197282 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2sl"] Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.271644 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-catalog-content\") pod \"redhat-marketplace-dl2sl\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.271994 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-utilities\") pod \"redhat-marketplace-dl2sl\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.272086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znjn7\" (UniqueName: \"kubernetes.io/projected/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-kube-api-access-znjn7\") pod \"redhat-marketplace-dl2sl\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.374584 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-catalog-content\") pod \"redhat-marketplace-dl2sl\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.374679 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-utilities\") pod \"redhat-marketplace-dl2sl\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.374793 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znjn7\" (UniqueName: \"kubernetes.io/projected/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-kube-api-access-znjn7\") pod \"redhat-marketplace-dl2sl\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.375221 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-catalog-content\") pod \"redhat-marketplace-dl2sl\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.375314 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-utilities\") pod \"redhat-marketplace-dl2sl\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.394811 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znjn7\" (UniqueName: \"kubernetes.io/projected/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-kube-api-access-znjn7\") pod \"redhat-marketplace-dl2sl\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.493222 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:19:52 crc kubenswrapper[4809]: I1127 18:19:52.943885 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2sl"] Nov 27 18:19:53 crc kubenswrapper[4809]: I1127 18:19:53.863480 4809 generic.go:334] "Generic (PLEG): container finished" podID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerID="20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537" exitCode=0 Nov 27 18:19:53 crc kubenswrapper[4809]: I1127 18:19:53.863585 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2sl" event={"ID":"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c","Type":"ContainerDied","Data":"20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537"} Nov 27 18:19:53 crc kubenswrapper[4809]: I1127 18:19:53.863853 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2sl" event={"ID":"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c","Type":"ContainerStarted","Data":"958d0420f4d6489570039baad8077fbabdef0925a2747511313cfd5bb44ac6d6"} Nov 27 18:19:55 crc kubenswrapper[4809]: I1127 18:19:55.900270 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2sl" event={"ID":"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c","Type":"ContainerStarted","Data":"10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe"} Nov 27 18:19:56 crc kubenswrapper[4809]: I1127 18:19:56.932228 4809 generic.go:334] "Generic (PLEG): container finished" podID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerID="10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe" exitCode=0 Nov 27 18:19:56 crc kubenswrapper[4809]: I1127 18:19:56.932617 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2sl" event={"ID":"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c","Type":"ContainerDied","Data":"10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe"} Nov 27 18:19:58 crc kubenswrapper[4809]: I1127 18:19:58.955361 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2sl" event={"ID":"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c","Type":"ContainerStarted","Data":"3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369"} Nov 27 18:19:58 crc kubenswrapper[4809]: I1127 18:19:58.983359 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dl2sl" podStartSLOduration=3.285426117 podStartE2EDuration="6.983337053s" podCreationTimestamp="2025-11-27 18:19:52 +0000 UTC" firstStartedPulling="2025-11-27 18:19:53.865493082 +0000 UTC m=+4229.137950434" lastFinishedPulling="2025-11-27 18:19:57.563404018 +0000 UTC m=+4232.835861370" observedRunningTime="2025-11-27 18:19:58.977908604 +0000 UTC m=+4234.250365956" watchObservedRunningTime="2025-11-27 18:19:58.983337053 +0000 UTC m=+4234.255794405" Nov 27 18:20:02 crc kubenswrapper[4809]: I1127 18:20:02.494263 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:20:02 crc kubenswrapper[4809]: I1127 18:20:02.495362 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:20:02 crc kubenswrapper[4809]: I1127 18:20:02.563828 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:20:12 crc kubenswrapper[4809]: I1127 18:20:12.546104 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:20:12 crc kubenswrapper[4809]: I1127 18:20:12.602243 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2sl"] Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.087596 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dl2sl" podUID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerName="registry-server" containerID="cri-o://3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369" gracePeriod=2 Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.547651 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.715842 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-utilities\") pod \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.715892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znjn7\" (UniqueName: \"kubernetes.io/projected/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-kube-api-access-znjn7\") pod \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.716012 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-catalog-content\") pod \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\" (UID: \"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c\") " Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.716795 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-utilities" (OuterVolumeSpecName: "utilities") pod "3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" (UID: "3ccb7dc6-5042-49aa-be1e-e594d3a43f1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.725399 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-kube-api-access-znjn7" (OuterVolumeSpecName: "kube-api-access-znjn7") pod "3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" (UID: "3ccb7dc6-5042-49aa-be1e-e594d3a43f1c"). InnerVolumeSpecName "kube-api-access-znjn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.762239 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" (UID: "3ccb7dc6-5042-49aa-be1e-e594d3a43f1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.818728 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.818800 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znjn7\" (UniqueName: \"kubernetes.io/projected/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-kube-api-access-znjn7\") on node \"crc\" DevicePath \"\"" Nov 27 18:20:13 crc kubenswrapper[4809]: I1127 18:20:13.818818 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.099055 4809 generic.go:334] "Generic (PLEG): container finished" podID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerID="3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369" exitCode=0 Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.099115 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2sl" event={"ID":"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c","Type":"ContainerDied","Data":"3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369"} Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.099167 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2sl" event={"ID":"3ccb7dc6-5042-49aa-be1e-e594d3a43f1c","Type":"ContainerDied","Data":"958d0420f4d6489570039baad8077fbabdef0925a2747511313cfd5bb44ac6d6"} Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.099191 4809 scope.go:117] "RemoveContainer" containerID="3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.099266 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dl2sl" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.139579 4809 scope.go:117] "RemoveContainer" containerID="10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.157689 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2sl"] Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.168944 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2sl"] Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.184895 4809 scope.go:117] "RemoveContainer" containerID="20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.228471 4809 scope.go:117] "RemoveContainer" containerID="3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369" Nov 27 18:20:14 crc kubenswrapper[4809]: E1127 18:20:14.228983 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369\": container with ID starting with 3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369 not found: ID does not exist" containerID="3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.229020 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369"} err="failed to get container status \"3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369\": rpc error: code = NotFound desc = could not find container \"3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369\": container with ID starting with 3bf1a8a4dd58ccf1f2158cc7369164631f27a53cd957edcadf300363ac85a369 not found: ID does not exist" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.229048 4809 scope.go:117] "RemoveContainer" containerID="10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe" Nov 27 18:20:14 crc kubenswrapper[4809]: E1127 18:20:14.229407 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe\": container with ID starting with 10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe not found: ID does not exist" containerID="10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.229445 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe"} err="failed to get container status \"10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe\": rpc error: code = NotFound desc = could not find container \"10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe\": container with ID starting with 10b80a657dcd9f26a0cc3368f936ddb6e8144a71731585937701efa8939e22fe not found: ID does not exist" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.229473 4809 scope.go:117] "RemoveContainer" containerID="20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537" Nov 27 18:20:14 crc kubenswrapper[4809]: E1127 18:20:14.229783 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537\": container with ID starting with 20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537 not found: ID does not exist" containerID="20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537" Nov 27 18:20:14 crc kubenswrapper[4809]: I1127 18:20:14.229804 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537"} err="failed to get container status \"20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537\": rpc error: code = NotFound desc = could not find container \"20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537\": container with ID starting with 20424e2a60ec0b2a9bfb8ee06e7a8a31160e0fe31f3c89a83eed5683dc131537 not found: ID does not exist" Nov 27 18:20:15 crc kubenswrapper[4809]: I1127 18:20:15.467967 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" path="/var/lib/kubelet/pods/3ccb7dc6-5042-49aa-be1e-e594d3a43f1c/volumes" Nov 27 18:20:42 crc kubenswrapper[4809]: I1127 18:20:42.982041 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f87td/must-gather-cxgjb"] Nov 27 18:20:42 crc kubenswrapper[4809]: E1127 18:20:42.983060 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerName="extract-utilities" Nov 27 18:20:42 crc kubenswrapper[4809]: I1127 18:20:42.983075 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerName="extract-utilities" Nov 27 18:20:42 crc kubenswrapper[4809]: E1127 18:20:42.983099 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerName="registry-server" Nov 27 18:20:42 crc kubenswrapper[4809]: I1127 18:20:42.983106 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerName="registry-server" Nov 27 18:20:42 crc kubenswrapper[4809]: E1127 18:20:42.983119 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerName="extract-content" Nov 27 18:20:42 crc kubenswrapper[4809]: I1127 18:20:42.983126 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerName="extract-content" Nov 27 18:20:42 crc kubenswrapper[4809]: I1127 18:20:42.983317 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ccb7dc6-5042-49aa-be1e-e594d3a43f1c" containerName="registry-server" Nov 27 18:20:42 crc kubenswrapper[4809]: I1127 18:20:42.984547 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:20:42 crc kubenswrapper[4809]: I1127 18:20:42.987277 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-f87td"/"openshift-service-ca.crt" Nov 27 18:20:42 crc kubenswrapper[4809]: I1127 18:20:42.991767 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-f87td"/"kube-root-ca.crt" Nov 27 18:20:42 crc kubenswrapper[4809]: I1127 18:20:42.993587 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-f87td/must-gather-cxgjb"] Nov 27 18:20:43 crc kubenswrapper[4809]: I1127 18:20:43.110671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90e34d76-f1e1-4304-aa3c-b789db90e515-must-gather-output\") pod \"must-gather-cxgjb\" (UID: \"90e34d76-f1e1-4304-aa3c-b789db90e515\") " pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:20:43 crc kubenswrapper[4809]: I1127 18:20:43.110752 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72wlm\" (UniqueName: \"kubernetes.io/projected/90e34d76-f1e1-4304-aa3c-b789db90e515-kube-api-access-72wlm\") pod \"must-gather-cxgjb\" (UID: \"90e34d76-f1e1-4304-aa3c-b789db90e515\") " pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:20:43 crc kubenswrapper[4809]: I1127 18:20:43.212287 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90e34d76-f1e1-4304-aa3c-b789db90e515-must-gather-output\") pod \"must-gather-cxgjb\" (UID: \"90e34d76-f1e1-4304-aa3c-b789db90e515\") " pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:20:43 crc kubenswrapper[4809]: I1127 18:20:43.212378 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72wlm\" (UniqueName: \"kubernetes.io/projected/90e34d76-f1e1-4304-aa3c-b789db90e515-kube-api-access-72wlm\") pod \"must-gather-cxgjb\" (UID: \"90e34d76-f1e1-4304-aa3c-b789db90e515\") " pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:20:43 crc kubenswrapper[4809]: I1127 18:20:43.212691 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90e34d76-f1e1-4304-aa3c-b789db90e515-must-gather-output\") pod \"must-gather-cxgjb\" (UID: \"90e34d76-f1e1-4304-aa3c-b789db90e515\") " pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:20:43 crc kubenswrapper[4809]: I1127 18:20:43.236557 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72wlm\" (UniqueName: \"kubernetes.io/projected/90e34d76-f1e1-4304-aa3c-b789db90e515-kube-api-access-72wlm\") pod \"must-gather-cxgjb\" (UID: \"90e34d76-f1e1-4304-aa3c-b789db90e515\") " pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:20:43 crc kubenswrapper[4809]: I1127 18:20:43.310584 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:20:43 crc kubenswrapper[4809]: I1127 18:20:43.746653 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-f87td/must-gather-cxgjb"] Nov 27 18:20:44 crc kubenswrapper[4809]: I1127 18:20:44.414438 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/must-gather-cxgjb" event={"ID":"90e34d76-f1e1-4304-aa3c-b789db90e515","Type":"ContainerStarted","Data":"4f138606d14f4db24113e14b880823000dfd3aad9ad52252458df42ba23748ac"} Nov 27 18:20:44 crc kubenswrapper[4809]: I1127 18:20:44.414881 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/must-gather-cxgjb" event={"ID":"90e34d76-f1e1-4304-aa3c-b789db90e515","Type":"ContainerStarted","Data":"e9ed1209ebfd9d503eaec1225afc556a8a0507c77eb4beedeec3fd6a85d45182"} Nov 27 18:20:44 crc kubenswrapper[4809]: I1127 18:20:44.414895 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/must-gather-cxgjb" event={"ID":"90e34d76-f1e1-4304-aa3c-b789db90e515","Type":"ContainerStarted","Data":"1fb67ece116bc5bd91785fd31575392fe9b3188dd04ee4a5156f20430c2b9ee2"} Nov 27 18:20:44 crc kubenswrapper[4809]: I1127 18:20:44.427526 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-f87td/must-gather-cxgjb" podStartSLOduration=2.427510989 podStartE2EDuration="2.427510989s" podCreationTimestamp="2025-11-27 18:20:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 18:20:44.426158013 +0000 UTC m=+4279.698615365" watchObservedRunningTime="2025-11-27 18:20:44.427510989 +0000 UTC m=+4279.699968331" Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.512156 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f87td/crc-debug-jlzhc"] Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.517981 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.520679 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-f87td"/"default-dockercfg-2n66d" Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.708094 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr9bh\" (UniqueName: \"kubernetes.io/projected/637ea4b5-d41a-4bbe-8304-1a435ab5a881-kube-api-access-sr9bh\") pod \"crc-debug-jlzhc\" (UID: \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\") " pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.708190 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/637ea4b5-d41a-4bbe-8304-1a435ab5a881-host\") pod \"crc-debug-jlzhc\" (UID: \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\") " pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.810273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr9bh\" (UniqueName: \"kubernetes.io/projected/637ea4b5-d41a-4bbe-8304-1a435ab5a881-kube-api-access-sr9bh\") pod \"crc-debug-jlzhc\" (UID: \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\") " pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.810811 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/637ea4b5-d41a-4bbe-8304-1a435ab5a881-host\") pod \"crc-debug-jlzhc\" (UID: \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\") " pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.811664 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/637ea4b5-d41a-4bbe-8304-1a435ab5a881-host\") pod \"crc-debug-jlzhc\" (UID: \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\") " pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.841432 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr9bh\" (UniqueName: \"kubernetes.io/projected/637ea4b5-d41a-4bbe-8304-1a435ab5a881-kube-api-access-sr9bh\") pod \"crc-debug-jlzhc\" (UID: \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\") " pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:20:47 crc kubenswrapper[4809]: I1127 18:20:47.842296 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:20:47 crc kubenswrapper[4809]: W1127 18:20:47.890460 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod637ea4b5_d41a_4bbe_8304_1a435ab5a881.slice/crio-8bee4f137ef52a7dc3552aabb79487c5dd17c66f251a70f1731bfd4c07c4d84d WatchSource:0}: Error finding container 8bee4f137ef52a7dc3552aabb79487c5dd17c66f251a70f1731bfd4c07c4d84d: Status 404 returned error can't find the container with id 8bee4f137ef52a7dc3552aabb79487c5dd17c66f251a70f1731bfd4c07c4d84d Nov 27 18:20:48 crc kubenswrapper[4809]: I1127 18:20:48.454820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/crc-debug-jlzhc" event={"ID":"637ea4b5-d41a-4bbe-8304-1a435ab5a881","Type":"ContainerStarted","Data":"2b51aafe8eecd1dd199c51ab68b09d5ce6f0f9a9397e27b67e2450366f79d5f8"} Nov 27 18:20:48 crc kubenswrapper[4809]: I1127 18:20:48.455580 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/crc-debug-jlzhc" event={"ID":"637ea4b5-d41a-4bbe-8304-1a435ab5a881","Type":"ContainerStarted","Data":"8bee4f137ef52a7dc3552aabb79487c5dd17c66f251a70f1731bfd4c07c4d84d"} Nov 27 18:20:48 crc kubenswrapper[4809]: I1127 18:20:48.471677 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-f87td/crc-debug-jlzhc" podStartSLOduration=1.471658035 podStartE2EDuration="1.471658035s" podCreationTimestamp="2025-11-27 18:20:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 18:20:48.469577938 +0000 UTC m=+4283.742035290" watchObservedRunningTime="2025-11-27 18:20:48.471658035 +0000 UTC m=+4283.744115387" Nov 27 18:20:55 crc kubenswrapper[4809]: I1127 18:20:55.779829 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:20:55 crc kubenswrapper[4809]: I1127 18:20:55.780264 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:21:25 crc kubenswrapper[4809]: I1127 18:21:25.776126 4809 generic.go:334] "Generic (PLEG): container finished" podID="637ea4b5-d41a-4bbe-8304-1a435ab5a881" containerID="2b51aafe8eecd1dd199c51ab68b09d5ce6f0f9a9397e27b67e2450366f79d5f8" exitCode=0 Nov 27 18:21:25 crc kubenswrapper[4809]: I1127 18:21:25.776205 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/crc-debug-jlzhc" event={"ID":"637ea4b5-d41a-4bbe-8304-1a435ab5a881","Type":"ContainerDied","Data":"2b51aafe8eecd1dd199c51ab68b09d5ce6f0f9a9397e27b67e2450366f79d5f8"} Nov 27 18:21:25 crc kubenswrapper[4809]: I1127 18:21:25.779664 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:21:25 crc kubenswrapper[4809]: I1127 18:21:25.779719 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:21:26 crc kubenswrapper[4809]: I1127 18:21:26.912077 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:21:26 crc kubenswrapper[4809]: I1127 18:21:26.948486 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f87td/crc-debug-jlzhc"] Nov 27 18:21:26 crc kubenswrapper[4809]: I1127 18:21:26.958804 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f87td/crc-debug-jlzhc"] Nov 27 18:21:26 crc kubenswrapper[4809]: I1127 18:21:26.989930 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr9bh\" (UniqueName: \"kubernetes.io/projected/637ea4b5-d41a-4bbe-8304-1a435ab5a881-kube-api-access-sr9bh\") pod \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\" (UID: \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\") " Nov 27 18:21:26 crc kubenswrapper[4809]: I1127 18:21:26.990257 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/637ea4b5-d41a-4bbe-8304-1a435ab5a881-host\") pod \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\" (UID: \"637ea4b5-d41a-4bbe-8304-1a435ab5a881\") " Nov 27 18:21:26 crc kubenswrapper[4809]: I1127 18:21:26.990371 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/637ea4b5-d41a-4bbe-8304-1a435ab5a881-host" (OuterVolumeSpecName: "host") pod "637ea4b5-d41a-4bbe-8304-1a435ab5a881" (UID: "637ea4b5-d41a-4bbe-8304-1a435ab5a881"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 18:21:26 crc kubenswrapper[4809]: I1127 18:21:26.990768 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/637ea4b5-d41a-4bbe-8304-1a435ab5a881-host\") on node \"crc\" DevicePath \"\"" Nov 27 18:21:26 crc kubenswrapper[4809]: I1127 18:21:26.995642 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/637ea4b5-d41a-4bbe-8304-1a435ab5a881-kube-api-access-sr9bh" (OuterVolumeSpecName: "kube-api-access-sr9bh") pod "637ea4b5-d41a-4bbe-8304-1a435ab5a881" (UID: "637ea4b5-d41a-4bbe-8304-1a435ab5a881"). InnerVolumeSpecName "kube-api-access-sr9bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:21:27 crc kubenswrapper[4809]: I1127 18:21:27.092563 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr9bh\" (UniqueName: \"kubernetes.io/projected/637ea4b5-d41a-4bbe-8304-1a435ab5a881-kube-api-access-sr9bh\") on node \"crc\" DevicePath \"\"" Nov 27 18:21:27 crc kubenswrapper[4809]: I1127 18:21:27.476495 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="637ea4b5-d41a-4bbe-8304-1a435ab5a881" path="/var/lib/kubelet/pods/637ea4b5-d41a-4bbe-8304-1a435ab5a881/volumes" Nov 27 18:21:27 crc kubenswrapper[4809]: E1127 18:21:27.687861 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod637ea4b5_d41a_4bbe_8304_1a435ab5a881.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod637ea4b5_d41a_4bbe_8304_1a435ab5a881.slice/crio-8bee4f137ef52a7dc3552aabb79487c5dd17c66f251a70f1731bfd4c07c4d84d\": RecentStats: unable to find data in memory cache]" Nov 27 18:21:27 crc kubenswrapper[4809]: I1127 18:21:27.795564 4809 scope.go:117] "RemoveContainer" containerID="2b51aafe8eecd1dd199c51ab68b09d5ce6f0f9a9397e27b67e2450366f79d5f8" Nov 27 18:21:27 crc kubenswrapper[4809]: I1127 18:21:27.795696 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-jlzhc" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.127092 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f87td/crc-debug-29g2l"] Nov 27 18:21:28 crc kubenswrapper[4809]: E1127 18:21:28.128095 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637ea4b5-d41a-4bbe-8304-1a435ab5a881" containerName="container-00" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.128117 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="637ea4b5-d41a-4bbe-8304-1a435ab5a881" containerName="container-00" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.128395 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="637ea4b5-d41a-4bbe-8304-1a435ab5a881" containerName="container-00" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.129351 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.131775 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-f87td"/"default-dockercfg-2n66d" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.223400 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fce1ea7-4c9b-4030-8497-107eba90263d-host\") pod \"crc-debug-29g2l\" (UID: \"9fce1ea7-4c9b-4030-8497-107eba90263d\") " pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.223599 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdxm7\" (UniqueName: \"kubernetes.io/projected/9fce1ea7-4c9b-4030-8497-107eba90263d-kube-api-access-bdxm7\") pod \"crc-debug-29g2l\" (UID: \"9fce1ea7-4c9b-4030-8497-107eba90263d\") " pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.326352 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdxm7\" (UniqueName: \"kubernetes.io/projected/9fce1ea7-4c9b-4030-8497-107eba90263d-kube-api-access-bdxm7\") pod \"crc-debug-29g2l\" (UID: \"9fce1ea7-4c9b-4030-8497-107eba90263d\") " pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.326481 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fce1ea7-4c9b-4030-8497-107eba90263d-host\") pod \"crc-debug-29g2l\" (UID: \"9fce1ea7-4c9b-4030-8497-107eba90263d\") " pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.326566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fce1ea7-4c9b-4030-8497-107eba90263d-host\") pod \"crc-debug-29g2l\" (UID: \"9fce1ea7-4c9b-4030-8497-107eba90263d\") " pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.349232 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdxm7\" (UniqueName: \"kubernetes.io/projected/9fce1ea7-4c9b-4030-8497-107eba90263d-kube-api-access-bdxm7\") pod \"crc-debug-29g2l\" (UID: \"9fce1ea7-4c9b-4030-8497-107eba90263d\") " pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.454534 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:28 crc kubenswrapper[4809]: W1127 18:21:28.479341 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fce1ea7_4c9b_4030_8497_107eba90263d.slice/crio-c967cdd38ef6662b6c7b389260fc86942205475d710b6da941034709baff25a2 WatchSource:0}: Error finding container c967cdd38ef6662b6c7b389260fc86942205475d710b6da941034709baff25a2: Status 404 returned error can't find the container with id c967cdd38ef6662b6c7b389260fc86942205475d710b6da941034709baff25a2 Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.817307 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/crc-debug-29g2l" event={"ID":"9fce1ea7-4c9b-4030-8497-107eba90263d","Type":"ContainerStarted","Data":"0a0b5fde061dbf9e60df761c23d28c250aaee03b37ee2692d9eac3de52ea7415"} Nov 27 18:21:28 crc kubenswrapper[4809]: I1127 18:21:28.817823 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/crc-debug-29g2l" event={"ID":"9fce1ea7-4c9b-4030-8497-107eba90263d","Type":"ContainerStarted","Data":"c967cdd38ef6662b6c7b389260fc86942205475d710b6da941034709baff25a2"} Nov 27 18:21:29 crc kubenswrapper[4809]: I1127 18:21:29.827874 4809 generic.go:334] "Generic (PLEG): container finished" podID="9fce1ea7-4c9b-4030-8497-107eba90263d" containerID="0a0b5fde061dbf9e60df761c23d28c250aaee03b37ee2692d9eac3de52ea7415" exitCode=0 Nov 27 18:21:29 crc kubenswrapper[4809]: I1127 18:21:29.828231 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/crc-debug-29g2l" event={"ID":"9fce1ea7-4c9b-4030-8497-107eba90263d","Type":"ContainerDied","Data":"0a0b5fde061dbf9e60df761c23d28c250aaee03b37ee2692d9eac3de52ea7415"} Nov 27 18:21:29 crc kubenswrapper[4809]: I1127 18:21:29.970295 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.053644 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fce1ea7-4c9b-4030-8497-107eba90263d-host\") pod \"9fce1ea7-4c9b-4030-8497-107eba90263d\" (UID: \"9fce1ea7-4c9b-4030-8497-107eba90263d\") " Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.053713 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fce1ea7-4c9b-4030-8497-107eba90263d-host" (OuterVolumeSpecName: "host") pod "9fce1ea7-4c9b-4030-8497-107eba90263d" (UID: "9fce1ea7-4c9b-4030-8497-107eba90263d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.053931 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdxm7\" (UniqueName: \"kubernetes.io/projected/9fce1ea7-4c9b-4030-8497-107eba90263d-kube-api-access-bdxm7\") pod \"9fce1ea7-4c9b-4030-8497-107eba90263d\" (UID: \"9fce1ea7-4c9b-4030-8497-107eba90263d\") " Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.054352 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fce1ea7-4c9b-4030-8497-107eba90263d-host\") on node \"crc\" DevicePath \"\"" Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.061346 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fce1ea7-4c9b-4030-8497-107eba90263d-kube-api-access-bdxm7" (OuterVolumeSpecName: "kube-api-access-bdxm7") pod "9fce1ea7-4c9b-4030-8497-107eba90263d" (UID: "9fce1ea7-4c9b-4030-8497-107eba90263d"). InnerVolumeSpecName "kube-api-access-bdxm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.102997 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f87td/crc-debug-29g2l"] Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.114545 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f87td/crc-debug-29g2l"] Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.157616 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdxm7\" (UniqueName: \"kubernetes.io/projected/9fce1ea7-4c9b-4030-8497-107eba90263d-kube-api-access-bdxm7\") on node \"crc\" DevicePath \"\"" Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.842664 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c967cdd38ef6662b6c7b389260fc86942205475d710b6da941034709baff25a2" Nov 27 18:21:30 crc kubenswrapper[4809]: I1127 18:21:30.843079 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-29g2l" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.308657 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f87td/crc-debug-47zmj"] Nov 27 18:21:31 crc kubenswrapper[4809]: E1127 18:21:31.309252 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fce1ea7-4c9b-4030-8497-107eba90263d" containerName="container-00" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.309280 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fce1ea7-4c9b-4030-8497-107eba90263d" containerName="container-00" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.309504 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fce1ea7-4c9b-4030-8497-107eba90263d" containerName="container-00" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.310358 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.312626 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-f87td"/"default-dockercfg-2n66d" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.383858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1033b10f-3499-4992-8f3a-85f06164f189-host\") pod \"crc-debug-47zmj\" (UID: \"1033b10f-3499-4992-8f3a-85f06164f189\") " pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.384364 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8tqx\" (UniqueName: \"kubernetes.io/projected/1033b10f-3499-4992-8f3a-85f06164f189-kube-api-access-c8tqx\") pod \"crc-debug-47zmj\" (UID: \"1033b10f-3499-4992-8f3a-85f06164f189\") " pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.491521 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8tqx\" (UniqueName: \"kubernetes.io/projected/1033b10f-3499-4992-8f3a-85f06164f189-kube-api-access-c8tqx\") pod \"crc-debug-47zmj\" (UID: \"1033b10f-3499-4992-8f3a-85f06164f189\") " pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.492198 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1033b10f-3499-4992-8f3a-85f06164f189-host\") pod \"crc-debug-47zmj\" (UID: \"1033b10f-3499-4992-8f3a-85f06164f189\") " pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.492362 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1033b10f-3499-4992-8f3a-85f06164f189-host\") pod \"crc-debug-47zmj\" (UID: \"1033b10f-3499-4992-8f3a-85f06164f189\") " pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.493180 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fce1ea7-4c9b-4030-8497-107eba90263d" path="/var/lib/kubelet/pods/9fce1ea7-4c9b-4030-8497-107eba90263d/volumes" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.511871 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8tqx\" (UniqueName: \"kubernetes.io/projected/1033b10f-3499-4992-8f3a-85f06164f189-kube-api-access-c8tqx\") pod \"crc-debug-47zmj\" (UID: \"1033b10f-3499-4992-8f3a-85f06164f189\") " pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.627901 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:31 crc kubenswrapper[4809]: W1127 18:21:31.655613 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1033b10f_3499_4992_8f3a_85f06164f189.slice/crio-c87cabf582aa927e351a11f3bf4720282d8d2fcc15cbfc342be6f8a27913d96c WatchSource:0}: Error finding container c87cabf582aa927e351a11f3bf4720282d8d2fcc15cbfc342be6f8a27913d96c: Status 404 returned error can't find the container with id c87cabf582aa927e351a11f3bf4720282d8d2fcc15cbfc342be6f8a27913d96c Nov 27 18:21:31 crc kubenswrapper[4809]: I1127 18:21:31.853175 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/crc-debug-47zmj" event={"ID":"1033b10f-3499-4992-8f3a-85f06164f189","Type":"ContainerStarted","Data":"c87cabf582aa927e351a11f3bf4720282d8d2fcc15cbfc342be6f8a27913d96c"} Nov 27 18:21:32 crc kubenswrapper[4809]: I1127 18:21:32.862205 4809 generic.go:334] "Generic (PLEG): container finished" podID="1033b10f-3499-4992-8f3a-85f06164f189" containerID="ffe47ae8bc84fafe2a73e401946ff98f3c07ce78acefcf032ae6c2bc87fe4487" exitCode=0 Nov 27 18:21:32 crc kubenswrapper[4809]: I1127 18:21:32.862254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/crc-debug-47zmj" event={"ID":"1033b10f-3499-4992-8f3a-85f06164f189","Type":"ContainerDied","Data":"ffe47ae8bc84fafe2a73e401946ff98f3c07ce78acefcf032ae6c2bc87fe4487"} Nov 27 18:21:32 crc kubenswrapper[4809]: I1127 18:21:32.897309 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f87td/crc-debug-47zmj"] Nov 27 18:21:32 crc kubenswrapper[4809]: I1127 18:21:32.908249 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f87td/crc-debug-47zmj"] Nov 27 18:21:33 crc kubenswrapper[4809]: I1127 18:21:33.979986 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:34 crc kubenswrapper[4809]: I1127 18:21:34.039352 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1033b10f-3499-4992-8f3a-85f06164f189-host\") pod \"1033b10f-3499-4992-8f3a-85f06164f189\" (UID: \"1033b10f-3499-4992-8f3a-85f06164f189\") " Nov 27 18:21:34 crc kubenswrapper[4809]: I1127 18:21:34.039489 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1033b10f-3499-4992-8f3a-85f06164f189-host" (OuterVolumeSpecName: "host") pod "1033b10f-3499-4992-8f3a-85f06164f189" (UID: "1033b10f-3499-4992-8f3a-85f06164f189"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 18:21:34 crc kubenswrapper[4809]: I1127 18:21:34.039549 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8tqx\" (UniqueName: \"kubernetes.io/projected/1033b10f-3499-4992-8f3a-85f06164f189-kube-api-access-c8tqx\") pod \"1033b10f-3499-4992-8f3a-85f06164f189\" (UID: \"1033b10f-3499-4992-8f3a-85f06164f189\") " Nov 27 18:21:34 crc kubenswrapper[4809]: I1127 18:21:34.040136 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1033b10f-3499-4992-8f3a-85f06164f189-host\") on node \"crc\" DevicePath \"\"" Nov 27 18:21:34 crc kubenswrapper[4809]: I1127 18:21:34.045260 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1033b10f-3499-4992-8f3a-85f06164f189-kube-api-access-c8tqx" (OuterVolumeSpecName: "kube-api-access-c8tqx") pod "1033b10f-3499-4992-8f3a-85f06164f189" (UID: "1033b10f-3499-4992-8f3a-85f06164f189"). InnerVolumeSpecName "kube-api-access-c8tqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:21:34 crc kubenswrapper[4809]: I1127 18:21:34.141542 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8tqx\" (UniqueName: \"kubernetes.io/projected/1033b10f-3499-4992-8f3a-85f06164f189-kube-api-access-c8tqx\") on node \"crc\" DevicePath \"\"" Nov 27 18:21:34 crc kubenswrapper[4809]: I1127 18:21:34.880991 4809 scope.go:117] "RemoveContainer" containerID="ffe47ae8bc84fafe2a73e401946ff98f3c07ce78acefcf032ae6c2bc87fe4487" Nov 27 18:21:34 crc kubenswrapper[4809]: I1127 18:21:34.881052 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/crc-debug-47zmj" Nov 27 18:21:35 crc kubenswrapper[4809]: I1127 18:21:35.468044 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1033b10f-3499-4992-8f3a-85f06164f189" path="/var/lib/kubelet/pods/1033b10f-3499-4992-8f3a-85f06164f189/volumes" Nov 27 18:21:53 crc kubenswrapper[4809]: I1127 18:21:53.791789 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5bfc74dc5d-snbcn_fea753ab-13fb-49aa-a430-6c5db50f1e6b/barbican-api/0.log" Nov 27 18:21:53 crc kubenswrapper[4809]: I1127 18:21:53.942049 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5bfc74dc5d-snbcn_fea753ab-13fb-49aa-a430-6c5db50f1e6b/barbican-api-log/0.log" Nov 27 18:21:53 crc kubenswrapper[4809]: I1127 18:21:53.953016 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-84d69cc6b6-vqlb7_9bc3b7a8-110d-481e-95d5-3d9034c0b2f9/barbican-keystone-listener/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.088907 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-84d69cc6b6-vqlb7_9bc3b7a8-110d-481e-95d5-3d9034c0b2f9/barbican-keystone-listener-log/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.148922 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-65b4758f4c-5krfp_af7083a1-8a26-49b0-8b7a-7b412592ed60/barbican-worker/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.162711 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-65b4758f4c-5krfp_af7083a1-8a26-49b0-8b7a-7b412592ed60/barbican-worker-log/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.339792 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-c9thq_1f21abe6-3443-4a72-b93b-574203c3f3b4/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.403363 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56782df1-096c-47b7-bebb-b0a0712a46cc/ceilometer-central-agent/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.538359 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56782df1-096c-47b7-bebb-b0a0712a46cc/ceilometer-notification-agent/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.547427 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56782df1-096c-47b7-bebb-b0a0712a46cc/sg-core/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.551078 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56782df1-096c-47b7-bebb-b0a0712a46cc/proxy-httpd/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.762981 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6/cinder-api-log/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.786776 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1adad9c7-0fbf-4ae3-ad2c-f3177a66b1c6/cinder-api/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.889047 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92/cinder-scheduler/0.log" Nov 27 18:21:54 crc kubenswrapper[4809]: I1127 18:21:54.961465 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5ea00edc-9b23-4bbe-93cc-cfae1ca2ff92/probe/0.log" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.030188 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-78ns4_2021472c-275a-44fd-802b-37e72374a72d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.172533 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-f8xvb_0db71fae-8a3a-4089-9ad8-e7a52bc5a6a3/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.264489 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-zgq88_0025bfa8-25eb-43cc-825e-c4e299f28d47/init/0.log" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.444882 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-zgq88_0025bfa8-25eb-43cc-825e-c4e299f28d47/init/0.log" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.455819 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-4t6mj_e5b77adb-7f43-4df6-b676-cb82003887fc/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.466035 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-zgq88_0025bfa8-25eb-43cc-825e-c4e299f28d47/dnsmasq-dns/0.log" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.779957 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.780020 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.780063 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.780591 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b000adc98c38ad552afdfcf0fbc8fde8da0963aa23c44e43a46c14ec901bccf"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 18:21:55 crc kubenswrapper[4809]: I1127 18:21:55.780647 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://7b000adc98c38ad552afdfcf0fbc8fde8da0963aa23c44e43a46c14ec901bccf" gracePeriod=600 Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.027270 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2339b0f8-362a-43b7-8679-8259e295d65e/glance-httpd/0.log" Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.028415 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2339b0f8-362a-43b7-8679-8259e295d65e/glance-log/0.log" Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.056975 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="7b000adc98c38ad552afdfcf0fbc8fde8da0963aa23c44e43a46c14ec901bccf" exitCode=0 Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.057026 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"7b000adc98c38ad552afdfcf0fbc8fde8da0963aa23c44e43a46c14ec901bccf"} Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.057056 4809 scope.go:117] "RemoveContainer" containerID="af3814217200ed1aa000d30e5b8edde0e8a63c41063428148278a320b8cf4ca5" Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.217156 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_78ba6922-17ec-497d-b15d-e2ba250a698e/glance-httpd/0.log" Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.241782 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_78ba6922-17ec-497d-b15d-e2ba250a698e/glance-log/0.log" Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.310365 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-qg9sr_43c7fb64-1bd5-4341-81e9-6129b4121106/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.490433 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-8wfrh_faa8dfb6-0171-4ad3-a68f-46ad6dd83646/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.888268 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29404441-hv6j5_47d4ccb6-b0c6-4676-a5d0-eb82e862457d/keystone-cron/0.log" Nov 27 18:21:56 crc kubenswrapper[4809]: I1127 18:21:56.977162 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-856cbdcf9c-lws7x_5ddd7431-85e1-4f2e-ae95-f00162cbe120/keystone-api/0.log" Nov 27 18:21:57 crc kubenswrapper[4809]: I1127 18:21:57.068413 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerStarted","Data":"5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779"} Nov 27 18:21:57 crc kubenswrapper[4809]: I1127 18:21:57.110028 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3ff647e6-3fcb-401c-a695-047f07ca9a58/kube-state-metrics/0.log" Nov 27 18:21:57 crc kubenswrapper[4809]: I1127 18:21:57.190268 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-2bdpp_80c21e8c-3128-48ef-91a8-365409103274/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:57 crc kubenswrapper[4809]: I1127 18:21:57.288244 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-ddw9s_565ebe47-4144-4730-9e10-ebd98010c9a3/logging-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:57 crc kubenswrapper[4809]: I1127 18:21:57.641201 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-65c6b6d897-rhmzw_1a7bab33-46e8-4e18-a0c2-1ca25e22493f/neutron-httpd/0.log" Nov 27 18:21:57 crc kubenswrapper[4809]: I1127 18:21:57.717573 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-65c6b6d897-rhmzw_1a7bab33-46e8-4e18-a0c2-1ca25e22493f/neutron-api/0.log" Nov 27 18:21:58 crc kubenswrapper[4809]: I1127 18:21:58.005878 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-74h2f_0d3439fb-cf21-47a9-bbf7-39a34a5469dd/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:58 crc kubenswrapper[4809]: I1127 18:21:58.547614 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_db48eb93-0220-4661-8de6-23aa1e43ca94/nova-api-log/0.log" Nov 27 18:21:58 crc kubenswrapper[4809]: I1127 18:21:58.630621 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_4fad2362-6b6a-45dd-8c49-9a5ef3a12f1b/nova-cell0-conductor-conductor/0.log" Nov 27 18:21:58 crc kubenswrapper[4809]: I1127 18:21:58.854691 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_01ecca97-66c5-4670-9304-5fb2b116e4be/nova-cell1-conductor-conductor/0.log" Nov 27 18:21:58 crc kubenswrapper[4809]: I1127 18:21:58.856302 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_db48eb93-0220-4661-8de6-23aa1e43ca94/nova-api-api/0.log" Nov 27 18:21:58 crc kubenswrapper[4809]: I1127 18:21:58.962632 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_969b1f1d-ce0b-48fd-b149-72649f406cf0/nova-cell1-novncproxy-novncproxy/0.log" Nov 27 18:21:59 crc kubenswrapper[4809]: I1127 18:21:59.120025 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-zc795_9230a20e-bbed-4c09-8e94-a87279b78723/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:21:59 crc kubenswrapper[4809]: I1127 18:21:59.293809 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_65be15f8-d10f-4079-b79a-eb92d2cdc432/nova-metadata-log/0.log" Nov 27 18:21:59 crc kubenswrapper[4809]: I1127 18:21:59.599554 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f82a4054-fea3-4071-b65d-529d900c0f04/mysql-bootstrap/0.log" Nov 27 18:21:59 crc kubenswrapper[4809]: I1127 18:21:59.637852 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f5bbe948-59f7-488a-b62f-ea8ead20541f/nova-scheduler-scheduler/0.log" Nov 27 18:21:59 crc kubenswrapper[4809]: I1127 18:21:59.762703 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f82a4054-fea3-4071-b65d-529d900c0f04/mysql-bootstrap/0.log" Nov 27 18:21:59 crc kubenswrapper[4809]: I1127 18:21:59.838403 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f82a4054-fea3-4071-b65d-529d900c0f04/galera/0.log" Nov 27 18:21:59 crc kubenswrapper[4809]: I1127 18:21:59.988487 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3cf12fa-2068-4bcd-ae42-d7c6dfea6782/mysql-bootstrap/0.log" Nov 27 18:22:00 crc kubenswrapper[4809]: I1127 18:22:00.227846 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3cf12fa-2068-4bcd-ae42-d7c6dfea6782/mysql-bootstrap/0.log" Nov 27 18:22:00 crc kubenswrapper[4809]: I1127 18:22:00.244588 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b3cf12fa-2068-4bcd-ae42-d7c6dfea6782/galera/0.log" Nov 27 18:22:00 crc kubenswrapper[4809]: I1127 18:22:00.426391 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_090560a3-12be-4a67-b507-0c7764053bf2/openstackclient/0.log" Nov 27 18:22:00 crc kubenswrapper[4809]: I1127 18:22:00.477288 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-h8g6t_55d50ebb-8cab-42df-96b5-9598262337a4/ovn-controller/0.log" Nov 27 18:22:00 crc kubenswrapper[4809]: I1127 18:22:00.530336 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_65be15f8-d10f-4079-b79a-eb92d2cdc432/nova-metadata-metadata/0.log" Nov 27 18:22:00 crc kubenswrapper[4809]: I1127 18:22:00.715794 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9567r_7a4f76ce-dff9-4d9c-a75b-7282bc630a36/openstack-network-exporter/0.log" Nov 27 18:22:00 crc kubenswrapper[4809]: I1127 18:22:00.731177 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jnvf9_a501e4bf-f184-451f-9b21-a813f147413d/ovsdb-server-init/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.010449 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jnvf9_a501e4bf-f184-451f-9b21-a813f147413d/ovsdb-server/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.020042 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jnvf9_a501e4bf-f184-451f-9b21-a813f147413d/ovsdb-server-init/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.039547 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jnvf9_a501e4bf-f184-451f-9b21-a813f147413d/ovs-vswitchd/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.242642 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-dwhp4_cc517a71-c16f-4144-94d8-36c2878d89d8/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.250977 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_545a6dbd-4f59-435b-b951-0a8e58be0d2a/openstack-network-exporter/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.283363 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_545a6dbd-4f59-435b-b951-0a8e58be0d2a/ovn-northd/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.432231 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6ab5ce4-ce26-4eee-98fe-1bb380df9376/openstack-network-exporter/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.515338 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6ab5ce4-ce26-4eee-98fe-1bb380df9376/ovsdbserver-nb/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.648329 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cda8dbab-96fd-4d1b-a9cb-e056382df1a2/openstack-network-exporter/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.706298 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cda8dbab-96fd-4d1b-a9cb-e056382df1a2/ovsdbserver-sb/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.844980 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-556f88f66d-dk28p_a43816b3-bfa7-4570-b082-9b217d660416/placement-api/0.log" Nov 27 18:22:01 crc kubenswrapper[4809]: I1127 18:22:01.949544 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-556f88f66d-dk28p_a43816b3-bfa7-4570-b082-9b217d660416/placement-log/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.025765 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3f0cad7c-85ca-4269-be47-555613716799/setup-container/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.203985 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3f0cad7c-85ca-4269-be47-555613716799/setup-container/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.241346 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3f0cad7c-85ca-4269-be47-555613716799/rabbitmq/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.248315 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_62edd287-129b-48b0-af61-d21bd27590d7/setup-container/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.471958 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_62edd287-129b-48b0-af61-d21bd27590d7/setup-container/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.536955 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_62edd287-129b-48b0-af61-d21bd27590d7/rabbitmq/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.557835 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-whjsn_562a0647-54b3-4b56-9717-24f2ab4ca87f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.734010 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-w5fcv_cfd74c64-0ab6-4fe3-a39f-18a6cd342d3e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.759316 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_bb517228-deb6-41fa-b6aa-354014cd34e1/memcached/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.783031 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ssrjm_56b0a916-b733-4a1c-b6dc-97925e319b83/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.919124 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-v55jx_5beb9875-497c-4eb8-9b9d-8474a6891b82/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:22:02 crc kubenswrapper[4809]: I1127 18:22:02.950491 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-wvz7x_7616f14b-3549-40b3-ba73-32b1cb830d98/ssh-known-hosts-edpm-deployment/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.112554 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5999bb96c5-jlmwd_2cdfac16-f905-422d-a059-60d8e3238bcf/proxy-server/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.155279 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-hgrfv_98c79518-5ef3-4382-a156-6de22fc92082/swift-ring-rebalance/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.211984 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5999bb96c5-jlmwd_2cdfac16-f905-422d-a059-60d8e3238bcf/proxy-httpd/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.322868 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/account-auditor/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.337056 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/account-reaper/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.370724 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/account-replicator/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.412525 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/account-server/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.452911 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/container-auditor/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.550361 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/container-server/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.575137 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/container-replicator/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.604418 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/container-updater/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.668209 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-auditor/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.696241 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-expirer/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.779946 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-replicator/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.786537 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-server/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.816815 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/object-updater/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.871351 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/rsync/0.log" Nov 27 18:22:03 crc kubenswrapper[4809]: I1127 18:22:03.924090 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1b359048-c31b-4d16-922a-77b4b9afe87d/swift-recon-cron/0.log" Nov 27 18:22:04 crc kubenswrapper[4809]: I1127 18:22:04.019339 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kv495_10c284aa-b6df-41bb-8ace-7dd18b804925/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:22:04 crc kubenswrapper[4809]: I1127 18:22:04.107119 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_79f6ceba-551a-427c-8690-1c8db833367d/tempest-tests-tempest-tests-runner/0.log" Nov 27 18:22:04 crc kubenswrapper[4809]: I1127 18:22:04.258198 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_a3526155-0e52-4f2c-b07a-920b8cf61639/test-operator-logs-container/0.log" Nov 27 18:22:04 crc kubenswrapper[4809]: I1127 18:22:04.303553 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-bz7g7_c91fe577-8e90-411e-b078-8ece235f2f93/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.370153 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/util/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.529037 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/util/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.557905 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/pull/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.583162 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/pull/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.717714 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/extract/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.725437 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/util/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.733251 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6f4c4c4addfe6a8c39cfe1d8e8f2248616a8d53b76fdcf42ead70ee287l6xjq_21484278-01d6-452c-806d-f8d364cb8325/pull/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.910572 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-g7pm5_4e5fbc76-cd41-432e-8c5b-90ff77d0704c/kube-rbac-proxy/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.924074 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-g7pm5_4e5fbc76-cd41-432e-8c5b-90ff77d0704c/manager/0.log" Nov 27 18:22:26 crc kubenswrapper[4809]: I1127 18:22:26.957451 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-swx59_be316447-6b83-4563-b75e-d4088882b695/kube-rbac-proxy/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.118079 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-swx59_be316447-6b83-4563-b75e-d4088882b695/manager/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.128992 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-l58tv_b470d686-15fd-4d31-ac6a-ffb60d3c9d79/kube-rbac-proxy/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.155346 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-l58tv_b470d686-15fd-4d31-ac6a-ffb60d3c9d79/manager/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.320232 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-x4fdl_f6774dea-1e4a-4cd0-8688-5a588fe6d49b/kube-rbac-proxy/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.398271 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-x4fdl_f6774dea-1e4a-4cd0-8688-5a588fe6d49b/manager/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.507601 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-4rlvc_27f6291c-27aa-430f-864e-05be868c89b9/kube-rbac-proxy/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.550951 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-4rlvc_27f6291c-27aa-430f-864e-05be868c89b9/manager/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.584961 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-mkngh_16fc4440-2960-4b7d-b059-c5192ea37a50/kube-rbac-proxy/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.734081 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-mkngh_16fc4440-2960-4b7d-b059-c5192ea37a50/manager/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.791722 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-b28r8_aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1/kube-rbac-proxy/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.943261 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-x9qq5_976fdd57-7e54-4896-abca-914b86a98b19/kube-rbac-proxy/0.log" Nov 27 18:22:27 crc kubenswrapper[4809]: I1127 18:22:27.963813 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-b28r8_aa9f8064-5a3b-4ed8-8c85-c77f54e22ff1/manager/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.057012 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-x9qq5_976fdd57-7e54-4896-abca-914b86a98b19/manager/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.113865 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-mwv49_63be6575-532b-487f-97a2-d9fd077e5be0/kube-rbac-proxy/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.179888 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-mwv49_63be6575-532b-487f-97a2-d9fd077e5be0/manager/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.293387 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-bzjd8_ad51c9a8-6543-403b-b8de-ff189da761a4/kube-rbac-proxy/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.335278 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-bzjd8_ad51c9a8-6543-403b-b8de-ff189da761a4/manager/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.477386 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-tcwh6_9fb9364a-e568-418b-b403-dfa071e60297/kube-rbac-proxy/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.496864 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-tcwh6_9fb9364a-e568-418b-b403-dfa071e60297/manager/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.545162 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-6t8sk_db1b7072-b244-4e32-ba19-b8cd4254af2f/kube-rbac-proxy/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.681055 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-6t8sk_db1b7072-b244-4e32-ba19-b8cd4254af2f/manager/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.725029 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-ll5l2_da239d06-b1c7-4b06-b830-5108030beed6/kube-rbac-proxy/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.854057 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-ll5l2_da239d06-b1c7-4b06-b830-5108030beed6/manager/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.908574 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-xrjzd_1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659/kube-rbac-proxy/0.log" Nov 27 18:22:28 crc kubenswrapper[4809]: I1127 18:22:28.956693 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-xrjzd_1ad10b2b-c3ec-4fa5-a828-1dfe7ce02659/manager/0.log" Nov 27 18:22:29 crc kubenswrapper[4809]: I1127 18:22:29.086028 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm_c077fc08-1ee2-493e-b631-f01ccd0b7c6e/kube-rbac-proxy/0.log" Nov 27 18:22:29 crc kubenswrapper[4809]: I1127 18:22:29.124919 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bwkwjm_c077fc08-1ee2-493e-b631-f01ccd0b7c6e/manager/0.log" Nov 27 18:22:29 crc kubenswrapper[4809]: I1127 18:22:29.519170 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-b44dff85c-jvvsd_cceba3cb-7f9b-49d8-96e4-3e0e3927d106/operator/0.log" Nov 27 18:22:29 crc kubenswrapper[4809]: I1127 18:22:29.553227 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-ckqts_45b7ef24-d97f-4dbe-8ad2-82f7f9983c68/registry-server/0.log" Nov 27 18:22:29 crc kubenswrapper[4809]: I1127 18:22:29.989751 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-c9pds_8ace4cdc-1253-4a0f-b667-83e997947f9b/kube-rbac-proxy/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.049785 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-c9pds_8ace4cdc-1253-4a0f-b667-83e997947f9b/manager/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.116396 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-8bq9l_3da94235-386c-4f05-b869-2e08cee40d5d/kube-rbac-proxy/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.233356 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-8bq9l_3da94235-386c-4f05-b869-2e08cee40d5d/manager/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.263490 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6644d5b8df-l9kv7_18ae1240-d812-4f11-b515-6e7a94b8468e/manager/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.351001 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5b9vp_21cb4ed4-198d-4dd0-b962-311bf2bcc850/operator/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.433091 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-vvtmf_8f8e02f4-c34c-4b0a-b18f-8e088d48e12c/kube-rbac-proxy/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.474167 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-vvtmf_8f8e02f4-c34c-4b0a-b18f-8e088d48e12c/manager/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.538972 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-ff79b6df5-lq72j_6016f85d-a0f4-45ee-a96f-95dcb2c57e3c/kube-rbac-proxy/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.599102 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-ff79b6df5-lq72j_6016f85d-a0f4-45ee-a96f-95dcb2c57e3c/manager/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.663909 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-j6nv4_8798a938-aa20-40e2-be26-408026fb123e/manager/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.674277 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-j6nv4_8798a938-aa20-40e2-be26-408026fb123e/kube-rbac-proxy/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.785035 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-d5hqw_5d49d329-6737-4bab-b243-64e618c0fa3b/kube-rbac-proxy/0.log" Nov 27 18:22:30 crc kubenswrapper[4809]: I1127 18:22:30.786048 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-d5hqw_5d49d329-6737-4bab-b243-64e618c0fa3b/manager/0.log" Nov 27 18:22:48 crc kubenswrapper[4809]: I1127 18:22:48.116566 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-tvl4x_87e2c04c-53e6-4524-90b8-b122e31b9280/control-plane-machine-set-operator/0.log" Nov 27 18:22:48 crc kubenswrapper[4809]: I1127 18:22:48.230537 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6rxcc_6d8b54b5-a20b-4461-bca8-c8ac2ec82048/kube-rbac-proxy/0.log" Nov 27 18:22:48 crc kubenswrapper[4809]: I1127 18:22:48.255694 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6rxcc_6d8b54b5-a20b-4461-bca8-c8ac2ec82048/machine-api-operator/0.log" Nov 27 18:23:00 crc kubenswrapper[4809]: I1127 18:23:00.908351 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7hzbp_df88093a-9522-45ca-913a-18a62128af63/cert-manager-controller/0.log" Nov 27 18:23:01 crc kubenswrapper[4809]: I1127 18:23:01.174311 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rczjv_04d521a8-9cbf-4ca8-b7a8-035504c16c17/cert-manager-cainjector/0.log" Nov 27 18:23:01 crc kubenswrapper[4809]: I1127 18:23:01.190704 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-tkpwp_79bf3153-df84-43fd-baa1-731fbc7f7873/cert-manager-webhook/0.log" Nov 27 18:23:14 crc kubenswrapper[4809]: I1127 18:23:14.005827 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-7lwv9_e0648dfc-fcc2-49c9-863b-0c40c33ea417/nmstate-console-plugin/0.log" Nov 27 18:23:14 crc kubenswrapper[4809]: I1127 18:23:14.183097 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-sccnl_acc7e45d-4ac0-43e7-a0ca-cd0ad3471f1b/nmstate-handler/0.log" Nov 27 18:23:14 crc kubenswrapper[4809]: I1127 18:23:14.211089 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-52jgm_637b5bb3-b183-43dd-bca4-0349ac11e7d5/kube-rbac-proxy/0.log" Nov 27 18:23:14 crc kubenswrapper[4809]: I1127 18:23:14.216102 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-52jgm_637b5bb3-b183-43dd-bca4-0349ac11e7d5/nmstate-metrics/0.log" Nov 27 18:23:14 crc kubenswrapper[4809]: I1127 18:23:14.776491 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-h86r5_796eb450-becc-4200-87e8-a9274d390901/nmstate-webhook/0.log" Nov 27 18:23:14 crc kubenswrapper[4809]: I1127 18:23:14.783584 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-pz6jf_aca17653-5316-4c78-8062-9526abf28d97/nmstate-operator/0.log" Nov 27 18:23:27 crc kubenswrapper[4809]: I1127 18:23:27.357005 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-fbb7c7d9f-bnjq7_2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61/manager/0.log" Nov 27 18:23:27 crc kubenswrapper[4809]: I1127 18:23:27.395042 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-fbb7c7d9f-bnjq7_2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61/kube-rbac-proxy/0.log" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.755527 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xd6sw"] Nov 27 18:23:31 crc kubenswrapper[4809]: E1127 18:23:31.757058 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1033b10f-3499-4992-8f3a-85f06164f189" containerName="container-00" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.757073 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1033b10f-3499-4992-8f3a-85f06164f189" containerName="container-00" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.757299 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1033b10f-3499-4992-8f3a-85f06164f189" containerName="container-00" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.758779 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.779072 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xd6sw"] Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.847655 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-utilities\") pod \"community-operators-xd6sw\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.847829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-catalog-content\") pod \"community-operators-xd6sw\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.848066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfb7s\" (UniqueName: \"kubernetes.io/projected/a9b99bf5-3d90-4d1b-becb-567518c743b7-kube-api-access-jfb7s\") pod \"community-operators-xd6sw\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.950189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-catalog-content\") pod \"community-operators-xd6sw\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.950632 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfb7s\" (UniqueName: \"kubernetes.io/projected/a9b99bf5-3d90-4d1b-becb-567518c743b7-kube-api-access-jfb7s\") pod \"community-operators-xd6sw\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.950684 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-utilities\") pod \"community-operators-xd6sw\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.950842 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-catalog-content\") pod \"community-operators-xd6sw\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.951077 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-utilities\") pod \"community-operators-xd6sw\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:31 crc kubenswrapper[4809]: I1127 18:23:31.968790 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfb7s\" (UniqueName: \"kubernetes.io/projected/a9b99bf5-3d90-4d1b-becb-567518c743b7-kube-api-access-jfb7s\") pod \"community-operators-xd6sw\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:32 crc kubenswrapper[4809]: I1127 18:23:32.077615 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:32 crc kubenswrapper[4809]: I1127 18:23:32.584387 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xd6sw"] Nov 27 18:23:32 crc kubenswrapper[4809]: I1127 18:23:32.924390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd6sw" event={"ID":"a9b99bf5-3d90-4d1b-becb-567518c743b7","Type":"ContainerStarted","Data":"58f2a3a1a8e283f0fb2b86e362240e85f5ff599ee8df8924f64219a075ae0fcb"} Nov 27 18:23:33 crc kubenswrapper[4809]: I1127 18:23:33.937143 4809 generic.go:334] "Generic (PLEG): container finished" podID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerID="4da465e2c9588c4a73ac824cbf6cb94f24413d2ca72fa728195c273bfa566aa2" exitCode=0 Nov 27 18:23:33 crc kubenswrapper[4809]: I1127 18:23:33.937234 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd6sw" event={"ID":"a9b99bf5-3d90-4d1b-becb-567518c743b7","Type":"ContainerDied","Data":"4da465e2c9588c4a73ac824cbf6cb94f24413d2ca72fa728195c273bfa566aa2"} Nov 27 18:23:33 crc kubenswrapper[4809]: I1127 18:23:33.939459 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 18:23:34 crc kubenswrapper[4809]: I1127 18:23:34.947894 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd6sw" event={"ID":"a9b99bf5-3d90-4d1b-becb-567518c743b7","Type":"ContainerStarted","Data":"4ca2ec8ffc724bd704e8bc1966521015e633fd1372e2e962e98fc2fd472d28ab"} Nov 27 18:23:35 crc kubenswrapper[4809]: I1127 18:23:35.958246 4809 generic.go:334] "Generic (PLEG): container finished" podID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerID="4ca2ec8ffc724bd704e8bc1966521015e633fd1372e2e962e98fc2fd472d28ab" exitCode=0 Nov 27 18:23:35 crc kubenswrapper[4809]: I1127 18:23:35.958318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd6sw" event={"ID":"a9b99bf5-3d90-4d1b-becb-567518c743b7","Type":"ContainerDied","Data":"4ca2ec8ffc724bd704e8bc1966521015e633fd1372e2e962e98fc2fd472d28ab"} Nov 27 18:23:36 crc kubenswrapper[4809]: I1127 18:23:36.971457 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd6sw" event={"ID":"a9b99bf5-3d90-4d1b-becb-567518c743b7","Type":"ContainerStarted","Data":"6b9e3e673583df011f522dd8df61de98bcbbf70791d37da60c03d2dff8729ab1"} Nov 27 18:23:36 crc kubenswrapper[4809]: I1127 18:23:36.986688 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xd6sw" podStartSLOduration=3.51669634 podStartE2EDuration="5.986668443s" podCreationTimestamp="2025-11-27 18:23:31 +0000 UTC" firstStartedPulling="2025-11-27 18:23:33.939063135 +0000 UTC m=+4449.211520487" lastFinishedPulling="2025-11-27 18:23:36.409035238 +0000 UTC m=+4451.681492590" observedRunningTime="2025-11-27 18:23:36.985523321 +0000 UTC m=+4452.257980673" watchObservedRunningTime="2025-11-27 18:23:36.986668443 +0000 UTC m=+4452.259125795" Nov 27 18:23:41 crc kubenswrapper[4809]: I1127 18:23:41.908753 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-8zwm2_c0697c42-98e3-49fd-b923-093806ec2876/cluster-logging-operator/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.066219 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-ndnhl_1cd5cf74-a3d1-42ab-a77d-3a6c0afff3c1/collector/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.078489 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.078542 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.106153 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_ed55809b-892d-4ba2-b037-5eb25f1add8d/loki-compactor/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.136143 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.278707 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-wmb6l_7d6e72ed-2edc-4744-8b78-adb9d95a41ee/loki-distributor/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.323317 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-86596f6498-4d4pc_4239a407-a58d-46ac-a13b-0f744c3fbf0f/gateway/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.444126 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-86596f6498-4d4pc_4239a407-a58d-46ac-a13b-0f744c3fbf0f/opa/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.509637 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-86596f6498-t5b62_42bade25-cca5-4a1b-9c2f-67c1ffb97c68/gateway/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.559910 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-86596f6498-t5b62_42bade25-cca5-4a1b-9c2f-67c1ffb97c68/opa/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.673838 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_cccadb3e-654a-430f-9c69-c99a6fde2279/loki-index-gateway/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.796494 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_87ecc35f-bd3c-43be-9841-0129299f73f9/loki-ingester/0.log" Nov 27 18:23:42 crc kubenswrapper[4809]: I1127 18:23:42.898281 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-gr498_ce25b13d-dcd3-420d-848a-afd09c1850f2/loki-querier/0.log" Nov 27 18:23:43 crc kubenswrapper[4809]: I1127 18:23:43.004550 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-flnsq_290cdb65-2424-4494-94cd-a69e5bbc126c/loki-query-frontend/0.log" Nov 27 18:23:43 crc kubenswrapper[4809]: I1127 18:23:43.089421 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:45 crc kubenswrapper[4809]: I1127 18:23:45.747544 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xd6sw"] Nov 27 18:23:45 crc kubenswrapper[4809]: I1127 18:23:45.748299 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xd6sw" podUID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerName="registry-server" containerID="cri-o://6b9e3e673583df011f522dd8df61de98bcbbf70791d37da60c03d2dff8729ab1" gracePeriod=2 Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.077057 4809 generic.go:334] "Generic (PLEG): container finished" podID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerID="6b9e3e673583df011f522dd8df61de98bcbbf70791d37da60c03d2dff8729ab1" exitCode=0 Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.077411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd6sw" event={"ID":"a9b99bf5-3d90-4d1b-becb-567518c743b7","Type":"ContainerDied","Data":"6b9e3e673583df011f522dd8df61de98bcbbf70791d37da60c03d2dff8729ab1"} Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.320372 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.445174 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-catalog-content\") pod \"a9b99bf5-3d90-4d1b-becb-567518c743b7\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.445244 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfb7s\" (UniqueName: \"kubernetes.io/projected/a9b99bf5-3d90-4d1b-becb-567518c743b7-kube-api-access-jfb7s\") pod \"a9b99bf5-3d90-4d1b-becb-567518c743b7\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.445762 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-utilities\") pod \"a9b99bf5-3d90-4d1b-becb-567518c743b7\" (UID: \"a9b99bf5-3d90-4d1b-becb-567518c743b7\") " Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.446329 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-utilities" (OuterVolumeSpecName: "utilities") pod "a9b99bf5-3d90-4d1b-becb-567518c743b7" (UID: "a9b99bf5-3d90-4d1b-becb-567518c743b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.446999 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.452514 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9b99bf5-3d90-4d1b-becb-567518c743b7-kube-api-access-jfb7s" (OuterVolumeSpecName: "kube-api-access-jfb7s") pod "a9b99bf5-3d90-4d1b-becb-567518c743b7" (UID: "a9b99bf5-3d90-4d1b-becb-567518c743b7"). InnerVolumeSpecName "kube-api-access-jfb7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.491371 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9b99bf5-3d90-4d1b-becb-567518c743b7" (UID: "a9b99bf5-3d90-4d1b-becb-567518c743b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.549027 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b99bf5-3d90-4d1b-becb-567518c743b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:23:46 crc kubenswrapper[4809]: I1127 18:23:46.549062 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfb7s\" (UniqueName: \"kubernetes.io/projected/a9b99bf5-3d90-4d1b-becb-567518c743b7-kube-api-access-jfb7s\") on node \"crc\" DevicePath \"\"" Nov 27 18:23:47 crc kubenswrapper[4809]: I1127 18:23:47.087941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd6sw" event={"ID":"a9b99bf5-3d90-4d1b-becb-567518c743b7","Type":"ContainerDied","Data":"58f2a3a1a8e283f0fb2b86e362240e85f5ff599ee8df8924f64219a075ae0fcb"} Nov 27 18:23:47 crc kubenswrapper[4809]: I1127 18:23:47.087997 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xd6sw" Nov 27 18:23:47 crc kubenswrapper[4809]: I1127 18:23:47.088315 4809 scope.go:117] "RemoveContainer" containerID="6b9e3e673583df011f522dd8df61de98bcbbf70791d37da60c03d2dff8729ab1" Nov 27 18:23:47 crc kubenswrapper[4809]: I1127 18:23:47.108011 4809 scope.go:117] "RemoveContainer" containerID="4ca2ec8ffc724bd704e8bc1966521015e633fd1372e2e962e98fc2fd472d28ab" Nov 27 18:23:47 crc kubenswrapper[4809]: I1127 18:23:47.123893 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xd6sw"] Nov 27 18:23:47 crc kubenswrapper[4809]: I1127 18:23:47.133804 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xd6sw"] Nov 27 18:23:47 crc kubenswrapper[4809]: I1127 18:23:47.140231 4809 scope.go:117] "RemoveContainer" containerID="4da465e2c9588c4a73ac824cbf6cb94f24413d2ca72fa728195c273bfa566aa2" Nov 27 18:23:47 crc kubenswrapper[4809]: I1127 18:23:47.469389 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9b99bf5-3d90-4d1b-becb-567518c743b7" path="/var/lib/kubelet/pods/a9b99bf5-3d90-4d1b-becb-567518c743b7/volumes" Nov 27 18:23:57 crc kubenswrapper[4809]: I1127 18:23:57.940640 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rbkql_976b64ef-1aff-4780-8353-dde2792907d4/kube-rbac-proxy/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.022112 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rbkql_976b64ef-1aff-4780-8353-dde2792907d4/controller/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.210702 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-frr-files/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.359165 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-reloader/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.404434 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-metrics/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.409237 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-frr-files/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.446103 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-reloader/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.593607 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-frr-files/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.633633 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-metrics/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.634332 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-metrics/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.646775 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-reloader/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.837161 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-reloader/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.846797 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-frr-files/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.865752 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/cp-metrics/0.log" Nov 27 18:23:58 crc kubenswrapper[4809]: I1127 18:23:58.898452 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/controller/0.log" Nov 27 18:23:59 crc kubenswrapper[4809]: I1127 18:23:59.022453 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/frr-metrics/0.log" Nov 27 18:23:59 crc kubenswrapper[4809]: I1127 18:23:59.074475 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/kube-rbac-proxy/0.log" Nov 27 18:23:59 crc kubenswrapper[4809]: I1127 18:23:59.140363 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/kube-rbac-proxy-frr/0.log" Nov 27 18:23:59 crc kubenswrapper[4809]: I1127 18:23:59.256074 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/reloader/0.log" Nov 27 18:23:59 crc kubenswrapper[4809]: I1127 18:23:59.386375 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-p48r2_42ffbf32-2946-4087-8474-c0c0ecead8eb/frr-k8s-webhook-server/0.log" Nov 27 18:23:59 crc kubenswrapper[4809]: I1127 18:23:59.604989 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-77cc889676-8xfh7_eca3efdb-8eaf-4bd0-a2a4-471180514ee5/manager/0.log" Nov 27 18:23:59 crc kubenswrapper[4809]: I1127 18:23:59.771202 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b59cf8cdf-wz5qf_779ed7ea-3841-4937-978f-046fe79f4a9c/webhook-server/0.log" Nov 27 18:23:59 crc kubenswrapper[4809]: I1127 18:23:59.914525 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-gwn22_283ae992-ff54-437d-aba9-3dc45f78d876/kube-rbac-proxy/0.log" Nov 27 18:24:00 crc kubenswrapper[4809]: I1127 18:24:00.545831 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-494tz_b7dfa097-6605-407b-8178-b4eb0497fd17/frr/0.log" Nov 27 18:24:00 crc kubenswrapper[4809]: I1127 18:24:00.553288 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-gwn22_283ae992-ff54-437d-aba9-3dc45f78d876/speaker/0.log" Nov 27 18:24:13 crc kubenswrapper[4809]: I1127 18:24:13.204379 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/util/0.log" Nov 27 18:24:13 crc kubenswrapper[4809]: I1127 18:24:13.357812 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/util/0.log" Nov 27 18:24:13 crc kubenswrapper[4809]: I1127 18:24:13.382330 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/pull/0.log" Nov 27 18:24:13 crc kubenswrapper[4809]: I1127 18:24:13.382504 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/pull/0.log" Nov 27 18:24:13 crc kubenswrapper[4809]: I1127 18:24:13.596905 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/util/0.log" Nov 27 18:24:13 crc kubenswrapper[4809]: I1127 18:24:13.613017 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/extract/0.log" Nov 27 18:24:13 crc kubenswrapper[4809]: I1127 18:24:13.613201 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6nxlj_96b8cee9-c201-410d-9391-cc02bb9a2eeb/pull/0.log" Nov 27 18:24:14 crc kubenswrapper[4809]: I1127 18:24:14.399463 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/util/0.log" Nov 27 18:24:14 crc kubenswrapper[4809]: I1127 18:24:14.538154 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/pull/0.log" Nov 27 18:24:14 crc kubenswrapper[4809]: I1127 18:24:14.569414 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/util/0.log" Nov 27 18:24:14 crc kubenswrapper[4809]: I1127 18:24:14.587595 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/pull/0.log" Nov 27 18:24:14 crc kubenswrapper[4809]: I1127 18:24:14.740314 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/util/0.log" Nov 27 18:24:14 crc kubenswrapper[4809]: I1127 18:24:14.743083 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/extract/0.log" Nov 27 18:24:14 crc kubenswrapper[4809]: I1127 18:24:14.777492 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83k68lf_11f39a67-042f-418d-9d4a-059243383431/pull/0.log" Nov 27 18:24:14 crc kubenswrapper[4809]: I1127 18:24:14.915663 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-utilities/0.log" Nov 27 18:24:15 crc kubenswrapper[4809]: I1127 18:24:15.091037 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-content/0.log" Nov 27 18:24:15 crc kubenswrapper[4809]: I1127 18:24:15.091780 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-content/0.log" Nov 27 18:24:15 crc kubenswrapper[4809]: I1127 18:24:15.119506 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-utilities/0.log" Nov 27 18:24:15 crc kubenswrapper[4809]: I1127 18:24:15.290386 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-content/0.log" Nov 27 18:24:15 crc kubenswrapper[4809]: I1127 18:24:15.336006 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/extract-utilities/0.log" Nov 27 18:24:15 crc kubenswrapper[4809]: I1127 18:24:15.889045 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ggq4r_aff3b1a0-ed26-4a23-b8fe-8ef38617c2d0/registry-server/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.165360 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-utilities/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.353682 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-content/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.363014 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-content/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.397262 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-utilities/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.556349 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-utilities/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.592896 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/extract-content/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.674232 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4x2bl_b3ffdbfe-1e8b-415f-9d97-47a17d84070c/marketplace-operator/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.813728 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-utilities/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.823545 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d4f9n_21562ef7-7977-4197-a0ff-c3ffc47e26e8/registry-server/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.985732 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-utilities/0.log" Nov 27 18:24:16 crc kubenswrapper[4809]: I1127 18:24:16.986491 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-content/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.010573 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-content/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.141276 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-utilities/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.162668 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/extract-content/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.248660 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-utilities/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.351052 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hk5rd_24a4f327-f4ee-45dd-938e-4de8c52b2e76/registry-server/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.586123 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-content/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.638701 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-utilities/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.656551 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-content/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.806903 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-content/0.log" Nov 27 18:24:17 crc kubenswrapper[4809]: I1127 18:24:17.826478 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/extract-utilities/0.log" Nov 27 18:24:18 crc kubenswrapper[4809]: I1127 18:24:18.360687 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8zfdv_ac0a6dfd-901d-47b5-aa9e-44e040d13e75/registry-server/0.log" Nov 27 18:24:25 crc kubenswrapper[4809]: I1127 18:24:25.779345 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:24:25 crc kubenswrapper[4809]: I1127 18:24:25.779916 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:24:42 crc kubenswrapper[4809]: I1127 18:24:42.784444 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-fbb7c7d9f-bnjq7_2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61/kube-rbac-proxy/0.log" Nov 27 18:24:42 crc kubenswrapper[4809]: I1127 18:24:42.813843 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-fbb7c7d9f-bnjq7_2f358eb9-3e6f-4efb-bd8f-6ef3320f3a61/manager/0.log" Nov 27 18:24:55 crc kubenswrapper[4809]: I1127 18:24:55.779692 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:24:55 crc kubenswrapper[4809]: I1127 18:24:55.780252 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:25:25 crc kubenswrapper[4809]: I1127 18:25:25.779534 4809 patch_prober.go:28] interesting pod/machine-config-daemon-qwx9w container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 18:25:25 crc kubenswrapper[4809]: I1127 18:25:25.780144 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 18:25:25 crc kubenswrapper[4809]: I1127 18:25:25.780189 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" Nov 27 18:25:25 crc kubenswrapper[4809]: I1127 18:25:25.781003 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779"} pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 18:25:25 crc kubenswrapper[4809]: I1127 18:25:25.781057 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerName="machine-config-daemon" containerID="cri-o://5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" gracePeriod=600 Nov 27 18:25:25 crc kubenswrapper[4809]: E1127 18:25:25.911344 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:25:26 crc kubenswrapper[4809]: I1127 18:25:26.019615 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" exitCode=0 Nov 27 18:25:26 crc kubenswrapper[4809]: I1127 18:25:26.019674 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" event={"ID":"6b63ded2-aa4f-4aba-b3cc-a3965a01036c","Type":"ContainerDied","Data":"5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779"} Nov 27 18:25:26 crc kubenswrapper[4809]: I1127 18:25:26.019719 4809 scope.go:117] "RemoveContainer" containerID="7b000adc98c38ad552afdfcf0fbc8fde8da0963aa23c44e43a46c14ec901bccf" Nov 27 18:25:26 crc kubenswrapper[4809]: I1127 18:25:26.020672 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:25:26 crc kubenswrapper[4809]: E1127 18:25:26.021109 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:25:37 crc kubenswrapper[4809]: I1127 18:25:37.463949 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:25:37 crc kubenswrapper[4809]: E1127 18:25:37.464632 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:25:52 crc kubenswrapper[4809]: I1127 18:25:52.458114 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:25:52 crc kubenswrapper[4809]: E1127 18:25:52.458944 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:26:07 crc kubenswrapper[4809]: I1127 18:26:07.458140 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:26:07 crc kubenswrapper[4809]: E1127 18:26:07.459021 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:26:16 crc kubenswrapper[4809]: I1127 18:26:16.519458 4809 generic.go:334] "Generic (PLEG): container finished" podID="90e34d76-f1e1-4304-aa3c-b789db90e515" containerID="e9ed1209ebfd9d503eaec1225afc556a8a0507c77eb4beedeec3fd6a85d45182" exitCode=0 Nov 27 18:26:16 crc kubenswrapper[4809]: I1127 18:26:16.519546 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f87td/must-gather-cxgjb" event={"ID":"90e34d76-f1e1-4304-aa3c-b789db90e515","Type":"ContainerDied","Data":"e9ed1209ebfd9d503eaec1225afc556a8a0507c77eb4beedeec3fd6a85d45182"} Nov 27 18:26:16 crc kubenswrapper[4809]: I1127 18:26:16.520803 4809 scope.go:117] "RemoveContainer" containerID="e9ed1209ebfd9d503eaec1225afc556a8a0507c77eb4beedeec3fd6a85d45182" Nov 27 18:26:17 crc kubenswrapper[4809]: I1127 18:26:17.508831 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f87td_must-gather-cxgjb_90e34d76-f1e1-4304-aa3c-b789db90e515/gather/0.log" Nov 27 18:26:22 crc kubenswrapper[4809]: I1127 18:26:22.458695 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:26:22 crc kubenswrapper[4809]: E1127 18:26:22.459640 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:26:28 crc kubenswrapper[4809]: I1127 18:26:28.508385 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f87td/must-gather-cxgjb"] Nov 27 18:26:28 crc kubenswrapper[4809]: I1127 18:26:28.510171 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-f87td/must-gather-cxgjb" podUID="90e34d76-f1e1-4304-aa3c-b789db90e515" containerName="copy" containerID="cri-o://4f138606d14f4db24113e14b880823000dfd3aad9ad52252458df42ba23748ac" gracePeriod=2 Nov 27 18:26:28 crc kubenswrapper[4809]: I1127 18:26:28.532310 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f87td/must-gather-cxgjb"] Nov 27 18:26:28 crc kubenswrapper[4809]: I1127 18:26:28.631021 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f87td_must-gather-cxgjb_90e34d76-f1e1-4304-aa3c-b789db90e515/copy/0.log" Nov 27 18:26:28 crc kubenswrapper[4809]: I1127 18:26:28.631636 4809 generic.go:334] "Generic (PLEG): container finished" podID="90e34d76-f1e1-4304-aa3c-b789db90e515" containerID="4f138606d14f4db24113e14b880823000dfd3aad9ad52252458df42ba23748ac" exitCode=143 Nov 27 18:26:28 crc kubenswrapper[4809]: I1127 18:26:28.931019 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f87td_must-gather-cxgjb_90e34d76-f1e1-4304-aa3c-b789db90e515/copy/0.log" Nov 27 18:26:28 crc kubenswrapper[4809]: I1127 18:26:28.931418 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.049988 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90e34d76-f1e1-4304-aa3c-b789db90e515-must-gather-output\") pod \"90e34d76-f1e1-4304-aa3c-b789db90e515\" (UID: \"90e34d76-f1e1-4304-aa3c-b789db90e515\") " Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.050371 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72wlm\" (UniqueName: \"kubernetes.io/projected/90e34d76-f1e1-4304-aa3c-b789db90e515-kube-api-access-72wlm\") pod \"90e34d76-f1e1-4304-aa3c-b789db90e515\" (UID: \"90e34d76-f1e1-4304-aa3c-b789db90e515\") " Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.061852 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90e34d76-f1e1-4304-aa3c-b789db90e515-kube-api-access-72wlm" (OuterVolumeSpecName: "kube-api-access-72wlm") pod "90e34d76-f1e1-4304-aa3c-b789db90e515" (UID: "90e34d76-f1e1-4304-aa3c-b789db90e515"). InnerVolumeSpecName "kube-api-access-72wlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.153003 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72wlm\" (UniqueName: \"kubernetes.io/projected/90e34d76-f1e1-4304-aa3c-b789db90e515-kube-api-access-72wlm\") on node \"crc\" DevicePath \"\"" Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.217189 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90e34d76-f1e1-4304-aa3c-b789db90e515-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "90e34d76-f1e1-4304-aa3c-b789db90e515" (UID: "90e34d76-f1e1-4304-aa3c-b789db90e515"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.255367 4809 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90e34d76-f1e1-4304-aa3c-b789db90e515-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.469071 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90e34d76-f1e1-4304-aa3c-b789db90e515" path="/var/lib/kubelet/pods/90e34d76-f1e1-4304-aa3c-b789db90e515/volumes" Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.642449 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f87td_must-gather-cxgjb_90e34d76-f1e1-4304-aa3c-b789db90e515/copy/0.log" Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.642940 4809 scope.go:117] "RemoveContainer" containerID="4f138606d14f4db24113e14b880823000dfd3aad9ad52252458df42ba23748ac" Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.643035 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f87td/must-gather-cxgjb" Nov 27 18:26:29 crc kubenswrapper[4809]: I1127 18:26:29.667719 4809 scope.go:117] "RemoveContainer" containerID="e9ed1209ebfd9d503eaec1225afc556a8a0507c77eb4beedeec3fd6a85d45182" Nov 27 18:26:37 crc kubenswrapper[4809]: I1127 18:26:37.458775 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:26:37 crc kubenswrapper[4809]: E1127 18:26:37.459808 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:26:50 crc kubenswrapper[4809]: I1127 18:26:50.458883 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:26:50 crc kubenswrapper[4809]: E1127 18:26:50.459625 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:27:01 crc kubenswrapper[4809]: I1127 18:27:01.458300 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:27:01 crc kubenswrapper[4809]: E1127 18:27:01.459222 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:27:13 crc kubenswrapper[4809]: I1127 18:27:13.458551 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:27:13 crc kubenswrapper[4809]: E1127 18:27:13.459527 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:27:25 crc kubenswrapper[4809]: I1127 18:27:25.464395 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:27:25 crc kubenswrapper[4809]: E1127 18:27:25.465229 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:27:35 crc kubenswrapper[4809]: I1127 18:27:35.977048 4809 scope.go:117] "RemoveContainer" containerID="0a0b5fde061dbf9e60df761c23d28c250aaee03b37ee2692d9eac3de52ea7415" Nov 27 18:27:38 crc kubenswrapper[4809]: I1127 18:27:38.458089 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:27:38 crc kubenswrapper[4809]: E1127 18:27:38.459688 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:27:51 crc kubenswrapper[4809]: I1127 18:27:51.460354 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:27:51 crc kubenswrapper[4809]: E1127 18:27:51.463315 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:28:05 crc kubenswrapper[4809]: I1127 18:28:05.464901 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:28:05 crc kubenswrapper[4809]: E1127 18:28:05.465838 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:28:16 crc kubenswrapper[4809]: I1127 18:28:16.458721 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:28:16 crc kubenswrapper[4809]: E1127 18:28:16.459592 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.602569 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cxznv"] Nov 27 18:28:19 crc kubenswrapper[4809]: E1127 18:28:19.603465 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e34d76-f1e1-4304-aa3c-b789db90e515" containerName="copy" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.603477 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e34d76-f1e1-4304-aa3c-b789db90e515" containerName="copy" Nov 27 18:28:19 crc kubenswrapper[4809]: E1127 18:28:19.603490 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerName="registry-server" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.603496 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerName="registry-server" Nov 27 18:28:19 crc kubenswrapper[4809]: E1127 18:28:19.603504 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerName="extract-content" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.603510 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerName="extract-content" Nov 27 18:28:19 crc kubenswrapper[4809]: E1127 18:28:19.603527 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e34d76-f1e1-4304-aa3c-b789db90e515" containerName="gather" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.603533 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e34d76-f1e1-4304-aa3c-b789db90e515" containerName="gather" Nov 27 18:28:19 crc kubenswrapper[4809]: E1127 18:28:19.603543 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerName="extract-utilities" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.603549 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerName="extract-utilities" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.603721 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e34d76-f1e1-4304-aa3c-b789db90e515" containerName="copy" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.603734 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9b99bf5-3d90-4d1b-becb-567518c743b7" containerName="registry-server" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.603767 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e34d76-f1e1-4304-aa3c-b789db90e515" containerName="gather" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.605167 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.619633 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cxznv"] Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.672786 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggmz4\" (UniqueName: \"kubernetes.io/projected/cd438bcb-7564-4cf0-84eb-3886f98d5672-kube-api-access-ggmz4\") pod \"redhat-operators-cxznv\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.672857 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-utilities\") pod \"redhat-operators-cxznv\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.672963 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-catalog-content\") pod \"redhat-operators-cxznv\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.774937 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggmz4\" (UniqueName: \"kubernetes.io/projected/cd438bcb-7564-4cf0-84eb-3886f98d5672-kube-api-access-ggmz4\") pod \"redhat-operators-cxznv\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.774978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-utilities\") pod \"redhat-operators-cxznv\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.775023 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-catalog-content\") pod \"redhat-operators-cxznv\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.775521 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-utilities\") pod \"redhat-operators-cxznv\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.775565 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-catalog-content\") pod \"redhat-operators-cxznv\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.795403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggmz4\" (UniqueName: \"kubernetes.io/projected/cd438bcb-7564-4cf0-84eb-3886f98d5672-kube-api-access-ggmz4\") pod \"redhat-operators-cxznv\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:19 crc kubenswrapper[4809]: I1127 18:28:19.938534 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:20 crc kubenswrapper[4809]: I1127 18:28:20.378367 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cxznv"] Nov 27 18:28:20 crc kubenswrapper[4809]: I1127 18:28:20.704187 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerID="e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff" exitCode=0 Nov 27 18:28:20 crc kubenswrapper[4809]: I1127 18:28:20.704252 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxznv" event={"ID":"cd438bcb-7564-4cf0-84eb-3886f98d5672","Type":"ContainerDied","Data":"e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff"} Nov 27 18:28:20 crc kubenswrapper[4809]: I1127 18:28:20.704299 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxznv" event={"ID":"cd438bcb-7564-4cf0-84eb-3886f98d5672","Type":"ContainerStarted","Data":"7f3aababd9c522741095c01de366da3d1750a9d033237a88f468ab20913e4b70"} Nov 27 18:28:22 crc kubenswrapper[4809]: I1127 18:28:22.726478 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxznv" event={"ID":"cd438bcb-7564-4cf0-84eb-3886f98d5672","Type":"ContainerStarted","Data":"00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84"} Nov 27 18:28:24 crc kubenswrapper[4809]: I1127 18:28:24.744185 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerID="00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84" exitCode=0 Nov 27 18:28:24 crc kubenswrapper[4809]: I1127 18:28:24.744262 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxznv" event={"ID":"cd438bcb-7564-4cf0-84eb-3886f98d5672","Type":"ContainerDied","Data":"00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84"} Nov 27 18:28:26 crc kubenswrapper[4809]: I1127 18:28:26.762893 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxznv" event={"ID":"cd438bcb-7564-4cf0-84eb-3886f98d5672","Type":"ContainerStarted","Data":"20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468"} Nov 27 18:28:26 crc kubenswrapper[4809]: I1127 18:28:26.781148 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cxznv" podStartSLOduration=2.619155277 podStartE2EDuration="7.781129062s" podCreationTimestamp="2025-11-27 18:28:19 +0000 UTC" firstStartedPulling="2025-11-27 18:28:20.706274812 +0000 UTC m=+4735.978732164" lastFinishedPulling="2025-11-27 18:28:25.868248597 +0000 UTC m=+4741.140705949" observedRunningTime="2025-11-27 18:28:26.779470866 +0000 UTC m=+4742.051928218" watchObservedRunningTime="2025-11-27 18:28:26.781129062 +0000 UTC m=+4742.053586414" Nov 27 18:28:29 crc kubenswrapper[4809]: I1127 18:28:29.939780 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:29 crc kubenswrapper[4809]: I1127 18:28:29.940891 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:30 crc kubenswrapper[4809]: I1127 18:28:30.989727 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cxznv" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerName="registry-server" probeResult="failure" output=< Nov 27 18:28:30 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Nov 27 18:28:30 crc kubenswrapper[4809]: > Nov 27 18:28:31 crc kubenswrapper[4809]: I1127 18:28:31.458185 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:28:31 crc kubenswrapper[4809]: E1127 18:28:31.458547 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:28:40 crc kubenswrapper[4809]: I1127 18:28:40.001175 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:40 crc kubenswrapper[4809]: I1127 18:28:40.061951 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:40 crc kubenswrapper[4809]: I1127 18:28:40.253883 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cxznv"] Nov 27 18:28:41 crc kubenswrapper[4809]: I1127 18:28:41.906188 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cxznv" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerName="registry-server" containerID="cri-o://20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468" gracePeriod=2 Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.346215 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.483933 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggmz4\" (UniqueName: \"kubernetes.io/projected/cd438bcb-7564-4cf0-84eb-3886f98d5672-kube-api-access-ggmz4\") pod \"cd438bcb-7564-4cf0-84eb-3886f98d5672\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.484325 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-catalog-content\") pod \"cd438bcb-7564-4cf0-84eb-3886f98d5672\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.484484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-utilities\") pod \"cd438bcb-7564-4cf0-84eb-3886f98d5672\" (UID: \"cd438bcb-7564-4cf0-84eb-3886f98d5672\") " Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.485662 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-utilities" (OuterVolumeSpecName: "utilities") pod "cd438bcb-7564-4cf0-84eb-3886f98d5672" (UID: "cd438bcb-7564-4cf0-84eb-3886f98d5672"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.489805 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd438bcb-7564-4cf0-84eb-3886f98d5672-kube-api-access-ggmz4" (OuterVolumeSpecName: "kube-api-access-ggmz4") pod "cd438bcb-7564-4cf0-84eb-3886f98d5672" (UID: "cd438bcb-7564-4cf0-84eb-3886f98d5672"). InnerVolumeSpecName "kube-api-access-ggmz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.588679 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.588716 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggmz4\" (UniqueName: \"kubernetes.io/projected/cd438bcb-7564-4cf0-84eb-3886f98d5672-kube-api-access-ggmz4\") on node \"crc\" DevicePath \"\"" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.597573 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd438bcb-7564-4cf0-84eb-3886f98d5672" (UID: "cd438bcb-7564-4cf0-84eb-3886f98d5672"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.690970 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd438bcb-7564-4cf0-84eb-3886f98d5672-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.919298 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerID="20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468" exitCode=0 Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.919387 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxznv" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.919404 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxznv" event={"ID":"cd438bcb-7564-4cf0-84eb-3886f98d5672","Type":"ContainerDied","Data":"20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468"} Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.920066 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxznv" event={"ID":"cd438bcb-7564-4cf0-84eb-3886f98d5672","Type":"ContainerDied","Data":"7f3aababd9c522741095c01de366da3d1750a9d033237a88f468ab20913e4b70"} Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.920087 4809 scope.go:117] "RemoveContainer" containerID="20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.945606 4809 scope.go:117] "RemoveContainer" containerID="00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84" Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.960027 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cxznv"] Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.969298 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cxznv"] Nov 27 18:28:42 crc kubenswrapper[4809]: I1127 18:28:42.989383 4809 scope.go:117] "RemoveContainer" containerID="e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff" Nov 27 18:28:43 crc kubenswrapper[4809]: I1127 18:28:43.014569 4809 scope.go:117] "RemoveContainer" containerID="20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468" Nov 27 18:28:43 crc kubenswrapper[4809]: E1127 18:28:43.014948 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468\": container with ID starting with 20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468 not found: ID does not exist" containerID="20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468" Nov 27 18:28:43 crc kubenswrapper[4809]: I1127 18:28:43.014993 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468"} err="failed to get container status \"20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468\": rpc error: code = NotFound desc = could not find container \"20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468\": container with ID starting with 20c39869b2f1e5702945fc057375164b01ba16ca3f85120f81f6599d7ad48468 not found: ID does not exist" Nov 27 18:28:43 crc kubenswrapper[4809]: I1127 18:28:43.015022 4809 scope.go:117] "RemoveContainer" containerID="00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84" Nov 27 18:28:43 crc kubenswrapper[4809]: E1127 18:28:43.015237 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84\": container with ID starting with 00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84 not found: ID does not exist" containerID="00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84" Nov 27 18:28:43 crc kubenswrapper[4809]: I1127 18:28:43.015256 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84"} err="failed to get container status \"00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84\": rpc error: code = NotFound desc = could not find container \"00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84\": container with ID starting with 00193525b5dcc0c5715fb85cb88bd8c9b1171b6924f797680b80345ca7b2af84 not found: ID does not exist" Nov 27 18:28:43 crc kubenswrapper[4809]: I1127 18:28:43.015267 4809 scope.go:117] "RemoveContainer" containerID="e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff" Nov 27 18:28:43 crc kubenswrapper[4809]: E1127 18:28:43.015526 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff\": container with ID starting with e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff not found: ID does not exist" containerID="e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff" Nov 27 18:28:43 crc kubenswrapper[4809]: I1127 18:28:43.015545 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff"} err="failed to get container status \"e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff\": rpc error: code = NotFound desc = could not find container \"e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff\": container with ID starting with e48f7b9b5ae73c0e657ff7f13ad4b8ac98798f8cf9c1f8a1573e97aa7f26beff not found: ID does not exist" Nov 27 18:28:43 crc kubenswrapper[4809]: I1127 18:28:43.458067 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:28:43 crc kubenswrapper[4809]: E1127 18:28:43.458644 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:28:43 crc kubenswrapper[4809]: I1127 18:28:43.476446 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" path="/var/lib/kubelet/pods/cd438bcb-7564-4cf0-84eb-3886f98d5672/volumes" Nov 27 18:28:57 crc kubenswrapper[4809]: I1127 18:28:57.459543 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:28:57 crc kubenswrapper[4809]: E1127 18:28:57.461596 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.245288 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8dgkz"] Nov 27 18:29:11 crc kubenswrapper[4809]: E1127 18:29:11.246486 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerName="extract-utilities" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.246506 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerName="extract-utilities" Nov 27 18:29:11 crc kubenswrapper[4809]: E1127 18:29:11.246526 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerName="extract-content" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.246535 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerName="extract-content" Nov 27 18:29:11 crc kubenswrapper[4809]: E1127 18:29:11.246555 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerName="registry-server" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.246566 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerName="registry-server" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.247127 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd438bcb-7564-4cf0-84eb-3886f98d5672" containerName="registry-server" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.249002 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.274423 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8dgkz"] Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.292602 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-catalog-content\") pod \"certified-operators-8dgkz\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.292805 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-utilities\") pod \"certified-operators-8dgkz\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.292862 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljts7\" (UniqueName: \"kubernetes.io/projected/ea3dc648-304d-430a-87a1-5a3c315ad238-kube-api-access-ljts7\") pod \"certified-operators-8dgkz\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.394809 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-utilities\") pod \"certified-operators-8dgkz\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.394872 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljts7\" (UniqueName: \"kubernetes.io/projected/ea3dc648-304d-430a-87a1-5a3c315ad238-kube-api-access-ljts7\") pod \"certified-operators-8dgkz\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.395403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-utilities\") pod \"certified-operators-8dgkz\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.395409 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-catalog-content\") pod \"certified-operators-8dgkz\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.404457 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-catalog-content\") pod \"certified-operators-8dgkz\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.427589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljts7\" (UniqueName: \"kubernetes.io/projected/ea3dc648-304d-430a-87a1-5a3c315ad238-kube-api-access-ljts7\") pod \"certified-operators-8dgkz\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:11 crc kubenswrapper[4809]: I1127 18:29:11.573206 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:12 crc kubenswrapper[4809]: I1127 18:29:12.043699 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8dgkz"] Nov 27 18:29:12 crc kubenswrapper[4809]: I1127 18:29:12.172173 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dgkz" event={"ID":"ea3dc648-304d-430a-87a1-5a3c315ad238","Type":"ContainerStarted","Data":"f4c188370caf5c64f579c376e81553e88967f5bc241aaf3205d1fb7c09b1a72b"} Nov 27 18:29:12 crc kubenswrapper[4809]: I1127 18:29:12.458026 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:29:12 crc kubenswrapper[4809]: E1127 18:29:12.458316 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:29:13 crc kubenswrapper[4809]: I1127 18:29:13.184446 4809 generic.go:334] "Generic (PLEG): container finished" podID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerID="2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c" exitCode=0 Nov 27 18:29:13 crc kubenswrapper[4809]: I1127 18:29:13.184500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dgkz" event={"ID":"ea3dc648-304d-430a-87a1-5a3c315ad238","Type":"ContainerDied","Data":"2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c"} Nov 27 18:29:13 crc kubenswrapper[4809]: I1127 18:29:13.188199 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 18:29:15 crc kubenswrapper[4809]: I1127 18:29:15.249282 4809 generic.go:334] "Generic (PLEG): container finished" podID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerID="cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5" exitCode=0 Nov 27 18:29:15 crc kubenswrapper[4809]: I1127 18:29:15.249478 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dgkz" event={"ID":"ea3dc648-304d-430a-87a1-5a3c315ad238","Type":"ContainerDied","Data":"cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5"} Nov 27 18:29:16 crc kubenswrapper[4809]: I1127 18:29:16.267363 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dgkz" event={"ID":"ea3dc648-304d-430a-87a1-5a3c315ad238","Type":"ContainerStarted","Data":"a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e"} Nov 27 18:29:16 crc kubenswrapper[4809]: I1127 18:29:16.302036 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8dgkz" podStartSLOduration=2.645280943 podStartE2EDuration="5.302017303s" podCreationTimestamp="2025-11-27 18:29:11 +0000 UTC" firstStartedPulling="2025-11-27 18:29:13.187972352 +0000 UTC m=+4788.460429704" lastFinishedPulling="2025-11-27 18:29:15.844708712 +0000 UTC m=+4791.117166064" observedRunningTime="2025-11-27 18:29:16.296929274 +0000 UTC m=+4791.569386626" watchObservedRunningTime="2025-11-27 18:29:16.302017303 +0000 UTC m=+4791.574474675" Nov 27 18:29:21 crc kubenswrapper[4809]: I1127 18:29:21.573841 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:21 crc kubenswrapper[4809]: I1127 18:29:21.574437 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:21 crc kubenswrapper[4809]: I1127 18:29:21.619353 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:22 crc kubenswrapper[4809]: I1127 18:29:22.361582 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:22 crc kubenswrapper[4809]: I1127 18:29:22.409487 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8dgkz"] Nov 27 18:29:23 crc kubenswrapper[4809]: I1127 18:29:23.458023 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:29:23 crc kubenswrapper[4809]: E1127 18:29:23.458330 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:29:24 crc kubenswrapper[4809]: I1127 18:29:24.341082 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8dgkz" podUID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerName="registry-server" containerID="cri-o://a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e" gracePeriod=2 Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.222251 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.292815 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-catalog-content\") pod \"ea3dc648-304d-430a-87a1-5a3c315ad238\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.292938 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljts7\" (UniqueName: \"kubernetes.io/projected/ea3dc648-304d-430a-87a1-5a3c315ad238-kube-api-access-ljts7\") pod \"ea3dc648-304d-430a-87a1-5a3c315ad238\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.293095 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-utilities\") pod \"ea3dc648-304d-430a-87a1-5a3c315ad238\" (UID: \"ea3dc648-304d-430a-87a1-5a3c315ad238\") " Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.293999 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-utilities" (OuterVolumeSpecName: "utilities") pod "ea3dc648-304d-430a-87a1-5a3c315ad238" (UID: "ea3dc648-304d-430a-87a1-5a3c315ad238"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.298168 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea3dc648-304d-430a-87a1-5a3c315ad238-kube-api-access-ljts7" (OuterVolumeSpecName: "kube-api-access-ljts7") pod "ea3dc648-304d-430a-87a1-5a3c315ad238" (UID: "ea3dc648-304d-430a-87a1-5a3c315ad238"). InnerVolumeSpecName "kube-api-access-ljts7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.341061 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea3dc648-304d-430a-87a1-5a3c315ad238" (UID: "ea3dc648-304d-430a-87a1-5a3c315ad238"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.351472 4809 generic.go:334] "Generic (PLEG): container finished" podID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerID="a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e" exitCode=0 Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.351519 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dgkz" event={"ID":"ea3dc648-304d-430a-87a1-5a3c315ad238","Type":"ContainerDied","Data":"a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e"} Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.351548 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dgkz" event={"ID":"ea3dc648-304d-430a-87a1-5a3c315ad238","Type":"ContainerDied","Data":"f4c188370caf5c64f579c376e81553e88967f5bc241aaf3205d1fb7c09b1a72b"} Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.351549 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8dgkz" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.351578 4809 scope.go:117] "RemoveContainer" containerID="a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.395291 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.395329 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea3dc648-304d-430a-87a1-5a3c315ad238-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.395340 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljts7\" (UniqueName: \"kubernetes.io/projected/ea3dc648-304d-430a-87a1-5a3c315ad238-kube-api-access-ljts7\") on node \"crc\" DevicePath \"\"" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.468096 4809 scope.go:117] "RemoveContainer" containerID="cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.490823 4809 scope.go:117] "RemoveContainer" containerID="2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.513725 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8dgkz"] Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.522120 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8dgkz"] Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.541553 4809 scope.go:117] "RemoveContainer" containerID="a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e" Nov 27 18:29:25 crc kubenswrapper[4809]: E1127 18:29:25.542000 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e\": container with ID starting with a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e not found: ID does not exist" containerID="a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.542038 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e"} err="failed to get container status \"a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e\": rpc error: code = NotFound desc = could not find container \"a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e\": container with ID starting with a62a7e90ab994f227f198bdd07ebf81b04c65e2cd5c6e15e32f1fa63ae31180e not found: ID does not exist" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.542059 4809 scope.go:117] "RemoveContainer" containerID="cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5" Nov 27 18:29:25 crc kubenswrapper[4809]: E1127 18:29:25.542390 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5\": container with ID starting with cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5 not found: ID does not exist" containerID="cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.542425 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5"} err="failed to get container status \"cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5\": rpc error: code = NotFound desc = could not find container \"cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5\": container with ID starting with cdddbe040e3d83dd25b1734592a83535e6f69bfe5e08a5e3d00c2ce0a8cba6b5 not found: ID does not exist" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.542447 4809 scope.go:117] "RemoveContainer" containerID="2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c" Nov 27 18:29:25 crc kubenswrapper[4809]: E1127 18:29:25.542817 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c\": container with ID starting with 2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c not found: ID does not exist" containerID="2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c" Nov 27 18:29:25 crc kubenswrapper[4809]: I1127 18:29:25.542847 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c"} err="failed to get container status \"2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c\": rpc error: code = NotFound desc = could not find container \"2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c\": container with ID starting with 2419029fc38f9e94ce07e5a3dc7d13191d952ce19341b0a822d65664800b2d1c not found: ID does not exist" Nov 27 18:29:27 crc kubenswrapper[4809]: I1127 18:29:27.468953 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea3dc648-304d-430a-87a1-5a3c315ad238" path="/var/lib/kubelet/pods/ea3dc648-304d-430a-87a1-5a3c315ad238/volumes" Nov 27 18:29:30 crc kubenswrapper[4809]: E1127 18:29:30.566755 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice/crio-f4c188370caf5c64f579c376e81553e88967f5bc241aaf3205d1fb7c09b1a72b\": RecentStats: unable to find data in memory cache]" Nov 27 18:29:38 crc kubenswrapper[4809]: I1127 18:29:38.458565 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:29:38 crc kubenswrapper[4809]: E1127 18:29:38.459360 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:29:40 crc kubenswrapper[4809]: E1127 18:29:40.798775 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice/crio-f4c188370caf5c64f579c376e81553e88967f5bc241aaf3205d1fb7c09b1a72b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice\": RecentStats: unable to find data in memory cache]" Nov 27 18:29:51 crc kubenswrapper[4809]: E1127 18:29:51.057421 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice/crio-f4c188370caf5c64f579c376e81553e88967f5bc241aaf3205d1fb7c09b1a72b\": RecentStats: unable to find data in memory cache]" Nov 27 18:29:52 crc kubenswrapper[4809]: I1127 18:29:52.457512 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:29:52 crc kubenswrapper[4809]: E1127 18:29:52.458265 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.175272 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv"] Nov 27 18:30:00 crc kubenswrapper[4809]: E1127 18:30:00.176385 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerName="extract-content" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.176439 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerName="extract-content" Nov 27 18:30:00 crc kubenswrapper[4809]: E1127 18:30:00.176458 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerName="registry-server" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.176468 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerName="registry-server" Nov 27 18:30:00 crc kubenswrapper[4809]: E1127 18:30:00.176488 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerName="extract-utilities" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.176496 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerName="extract-utilities" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.176791 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea3dc648-304d-430a-87a1-5a3c315ad238" containerName="registry-server" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.177671 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.184993 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.185131 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.192513 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv"] Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.219972 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2d44070-ab37-46c6-9713-d4f4853338f0-secret-volume\") pod \"collect-profiles-29404470-8swdv\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.220106 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2d44070-ab37-46c6-9713-d4f4853338f0-config-volume\") pod \"collect-profiles-29404470-8swdv\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.321840 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2d44070-ab37-46c6-9713-d4f4853338f0-secret-volume\") pod \"collect-profiles-29404470-8swdv\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.321914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql4cs\" (UniqueName: \"kubernetes.io/projected/f2d44070-ab37-46c6-9713-d4f4853338f0-kube-api-access-ql4cs\") pod \"collect-profiles-29404470-8swdv\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.321967 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2d44070-ab37-46c6-9713-d4f4853338f0-config-volume\") pod \"collect-profiles-29404470-8swdv\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.322855 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2d44070-ab37-46c6-9713-d4f4853338f0-config-volume\") pod \"collect-profiles-29404470-8swdv\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.329643 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2d44070-ab37-46c6-9713-d4f4853338f0-secret-volume\") pod \"collect-profiles-29404470-8swdv\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.424153 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql4cs\" (UniqueName: \"kubernetes.io/projected/f2d44070-ab37-46c6-9713-d4f4853338f0-kube-api-access-ql4cs\") pod \"collect-profiles-29404470-8swdv\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.440831 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql4cs\" (UniqueName: \"kubernetes.io/projected/f2d44070-ab37-46c6-9713-d4f4853338f0-kube-api-access-ql4cs\") pod \"collect-profiles-29404470-8swdv\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.512849 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:00 crc kubenswrapper[4809]: I1127 18:30:00.947600 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv"] Nov 27 18:30:01 crc kubenswrapper[4809]: E1127 18:30:01.325559 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice/crio-f4c188370caf5c64f579c376e81553e88967f5bc241aaf3205d1fb7c09b1a72b\": RecentStats: unable to find data in memory cache]" Nov 27 18:30:01 crc kubenswrapper[4809]: I1127 18:30:01.663284 4809 generic.go:334] "Generic (PLEG): container finished" podID="f2d44070-ab37-46c6-9713-d4f4853338f0" containerID="8cbe8e256aeec367974eec4b55bf8191f410eab21c6258edc97cdc996deb288a" exitCode=0 Nov 27 18:30:01 crc kubenswrapper[4809]: I1127 18:30:01.663420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" event={"ID":"f2d44070-ab37-46c6-9713-d4f4853338f0","Type":"ContainerDied","Data":"8cbe8e256aeec367974eec4b55bf8191f410eab21c6258edc97cdc996deb288a"} Nov 27 18:30:01 crc kubenswrapper[4809]: I1127 18:30:01.663612 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" event={"ID":"f2d44070-ab37-46c6-9713-d4f4853338f0","Type":"ContainerStarted","Data":"85ce64d0d880afa2b03b0ada4c11c878efb31d52b856e167762e4f29af17fe8c"} Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.083681 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.178114 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql4cs\" (UniqueName: \"kubernetes.io/projected/f2d44070-ab37-46c6-9713-d4f4853338f0-kube-api-access-ql4cs\") pod \"f2d44070-ab37-46c6-9713-d4f4853338f0\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.178176 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2d44070-ab37-46c6-9713-d4f4853338f0-secret-volume\") pod \"f2d44070-ab37-46c6-9713-d4f4853338f0\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.178341 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2d44070-ab37-46c6-9713-d4f4853338f0-config-volume\") pod \"f2d44070-ab37-46c6-9713-d4f4853338f0\" (UID: \"f2d44070-ab37-46c6-9713-d4f4853338f0\") " Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.179599 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2d44070-ab37-46c6-9713-d4f4853338f0-config-volume" (OuterVolumeSpecName: "config-volume") pod "f2d44070-ab37-46c6-9713-d4f4853338f0" (UID: "f2d44070-ab37-46c6-9713-d4f4853338f0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.185702 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2d44070-ab37-46c6-9713-d4f4853338f0-kube-api-access-ql4cs" (OuterVolumeSpecName: "kube-api-access-ql4cs") pod "f2d44070-ab37-46c6-9713-d4f4853338f0" (UID: "f2d44070-ab37-46c6-9713-d4f4853338f0"). InnerVolumeSpecName "kube-api-access-ql4cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.185766 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d44070-ab37-46c6-9713-d4f4853338f0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f2d44070-ab37-46c6-9713-d4f4853338f0" (UID: "f2d44070-ab37-46c6-9713-d4f4853338f0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.280415 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql4cs\" (UniqueName: \"kubernetes.io/projected/f2d44070-ab37-46c6-9713-d4f4853338f0-kube-api-access-ql4cs\") on node \"crc\" DevicePath \"\"" Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.280688 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2d44070-ab37-46c6-9713-d4f4853338f0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.280790 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2d44070-ab37-46c6-9713-d4f4853338f0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.680478 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" event={"ID":"f2d44070-ab37-46c6-9713-d4f4853338f0","Type":"ContainerDied","Data":"85ce64d0d880afa2b03b0ada4c11c878efb31d52b856e167762e4f29af17fe8c"} Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.680516 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404470-8swdv" Nov 27 18:30:03 crc kubenswrapper[4809]: I1127 18:30:03.680527 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85ce64d0d880afa2b03b0ada4c11c878efb31d52b856e167762e4f29af17fe8c" Nov 27 18:30:04 crc kubenswrapper[4809]: I1127 18:30:04.158432 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d"] Nov 27 18:30:04 crc kubenswrapper[4809]: I1127 18:30:04.167116 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404425-j5p8d"] Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.230467 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2gc"] Nov 27 18:30:05 crc kubenswrapper[4809]: E1127 18:30:05.230885 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2d44070-ab37-46c6-9713-d4f4853338f0" containerName="collect-profiles" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.230896 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2d44070-ab37-46c6-9713-d4f4853338f0" containerName="collect-profiles" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.231111 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2d44070-ab37-46c6-9713-d4f4853338f0" containerName="collect-profiles" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.232463 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.245665 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2gc"] Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.320103 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-utilities\") pod \"redhat-marketplace-dl2gc\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.320282 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-catalog-content\") pod \"redhat-marketplace-dl2gc\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.320426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk4p2\" (UniqueName: \"kubernetes.io/projected/006fbfe2-b2e2-4707-aa60-898e8feb6e76-kube-api-access-nk4p2\") pod \"redhat-marketplace-dl2gc\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.421881 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk4p2\" (UniqueName: \"kubernetes.io/projected/006fbfe2-b2e2-4707-aa60-898e8feb6e76-kube-api-access-nk4p2\") pod \"redhat-marketplace-dl2gc\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.421993 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-utilities\") pod \"redhat-marketplace-dl2gc\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.422071 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-catalog-content\") pod \"redhat-marketplace-dl2gc\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.422589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-utilities\") pod \"redhat-marketplace-dl2gc\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.422608 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-catalog-content\") pod \"redhat-marketplace-dl2gc\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.443178 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk4p2\" (UniqueName: \"kubernetes.io/projected/006fbfe2-b2e2-4707-aa60-898e8feb6e76-kube-api-access-nk4p2\") pod \"redhat-marketplace-dl2gc\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.470142 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5428a8-1555-48ba-9b96-836b994671c0" path="/var/lib/kubelet/pods/8a5428a8-1555-48ba-9b96-836b994671c0/volumes" Nov 27 18:30:05 crc kubenswrapper[4809]: I1127 18:30:05.562833 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:06 crc kubenswrapper[4809]: I1127 18:30:06.025403 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2gc"] Nov 27 18:30:06 crc kubenswrapper[4809]: I1127 18:30:06.458558 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:30:06 crc kubenswrapper[4809]: E1127 18:30:06.459210 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:30:06 crc kubenswrapper[4809]: I1127 18:30:06.713163 4809 generic.go:334] "Generic (PLEG): container finished" podID="006fbfe2-b2e2-4707-aa60-898e8feb6e76" containerID="7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482" exitCode=0 Nov 27 18:30:06 crc kubenswrapper[4809]: I1127 18:30:06.713247 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2gc" event={"ID":"006fbfe2-b2e2-4707-aa60-898e8feb6e76","Type":"ContainerDied","Data":"7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482"} Nov 27 18:30:06 crc kubenswrapper[4809]: I1127 18:30:06.713320 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2gc" event={"ID":"006fbfe2-b2e2-4707-aa60-898e8feb6e76","Type":"ContainerStarted","Data":"d12c01081be1c1e6e262d1a576e9d2bf7d74369e1d57632323a66c4de0a228fb"} Nov 27 18:30:07 crc kubenswrapper[4809]: I1127 18:30:07.725309 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2gc" event={"ID":"006fbfe2-b2e2-4707-aa60-898e8feb6e76","Type":"ContainerStarted","Data":"a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6"} Nov 27 18:30:08 crc kubenswrapper[4809]: I1127 18:30:08.734783 4809 generic.go:334] "Generic (PLEG): container finished" podID="006fbfe2-b2e2-4707-aa60-898e8feb6e76" containerID="a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6" exitCode=0 Nov 27 18:30:08 crc kubenswrapper[4809]: I1127 18:30:08.734852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2gc" event={"ID":"006fbfe2-b2e2-4707-aa60-898e8feb6e76","Type":"ContainerDied","Data":"a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6"} Nov 27 18:30:09 crc kubenswrapper[4809]: I1127 18:30:09.747984 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2gc" event={"ID":"006fbfe2-b2e2-4707-aa60-898e8feb6e76","Type":"ContainerStarted","Data":"3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4"} Nov 27 18:30:09 crc kubenswrapper[4809]: I1127 18:30:09.771063 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dl2gc" podStartSLOduration=2.249667926 podStartE2EDuration="4.771029981s" podCreationTimestamp="2025-11-27 18:30:05 +0000 UTC" firstStartedPulling="2025-11-27 18:30:06.715327192 +0000 UTC m=+4841.987784544" lastFinishedPulling="2025-11-27 18:30:09.236689247 +0000 UTC m=+4844.509146599" observedRunningTime="2025-11-27 18:30:09.770809665 +0000 UTC m=+4845.043267017" watchObservedRunningTime="2025-11-27 18:30:09.771029981 +0000 UTC m=+4845.043487333" Nov 27 18:30:11 crc kubenswrapper[4809]: E1127 18:30:11.567411 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice/crio-f4c188370caf5c64f579c376e81553e88967f5bc241aaf3205d1fb7c09b1a72b\": RecentStats: unable to find data in memory cache]" Nov 27 18:30:15 crc kubenswrapper[4809]: I1127 18:30:15.563774 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:15 crc kubenswrapper[4809]: I1127 18:30:15.565110 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:15 crc kubenswrapper[4809]: I1127 18:30:15.623493 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:15 crc kubenswrapper[4809]: I1127 18:30:15.843518 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:15 crc kubenswrapper[4809]: I1127 18:30:15.887275 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2gc"] Nov 27 18:30:17 crc kubenswrapper[4809]: I1127 18:30:17.815789 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dl2gc" podUID="006fbfe2-b2e2-4707-aa60-898e8feb6e76" containerName="registry-server" containerID="cri-o://3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4" gracePeriod=2 Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.307089 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.370370 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk4p2\" (UniqueName: \"kubernetes.io/projected/006fbfe2-b2e2-4707-aa60-898e8feb6e76-kube-api-access-nk4p2\") pod \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.370421 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-catalog-content\") pod \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.370689 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-utilities\") pod \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\" (UID: \"006fbfe2-b2e2-4707-aa60-898e8feb6e76\") " Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.372058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-utilities" (OuterVolumeSpecName: "utilities") pod "006fbfe2-b2e2-4707-aa60-898e8feb6e76" (UID: "006fbfe2-b2e2-4707-aa60-898e8feb6e76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.379229 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/006fbfe2-b2e2-4707-aa60-898e8feb6e76-kube-api-access-nk4p2" (OuterVolumeSpecName: "kube-api-access-nk4p2") pod "006fbfe2-b2e2-4707-aa60-898e8feb6e76" (UID: "006fbfe2-b2e2-4707-aa60-898e8feb6e76"). InnerVolumeSpecName "kube-api-access-nk4p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.393661 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "006fbfe2-b2e2-4707-aa60-898e8feb6e76" (UID: "006fbfe2-b2e2-4707-aa60-898e8feb6e76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.473870 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.473905 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk4p2\" (UniqueName: \"kubernetes.io/projected/006fbfe2-b2e2-4707-aa60-898e8feb6e76-kube-api-access-nk4p2\") on node \"crc\" DevicePath \"\"" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.473954 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006fbfe2-b2e2-4707-aa60-898e8feb6e76-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.825569 4809 generic.go:334] "Generic (PLEG): container finished" podID="006fbfe2-b2e2-4707-aa60-898e8feb6e76" containerID="3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4" exitCode=0 Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.825610 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2gc" event={"ID":"006fbfe2-b2e2-4707-aa60-898e8feb6e76","Type":"ContainerDied","Data":"3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4"} Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.825637 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dl2gc" event={"ID":"006fbfe2-b2e2-4707-aa60-898e8feb6e76","Type":"ContainerDied","Data":"d12c01081be1c1e6e262d1a576e9d2bf7d74369e1d57632323a66c4de0a228fb"} Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.825646 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dl2gc" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.825658 4809 scope.go:117] "RemoveContainer" containerID="3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.844298 4809 scope.go:117] "RemoveContainer" containerID="a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.861141 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2gc"] Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.870157 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dl2gc"] Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.879153 4809 scope.go:117] "RemoveContainer" containerID="7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.917884 4809 scope.go:117] "RemoveContainer" containerID="3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4" Nov 27 18:30:18 crc kubenswrapper[4809]: E1127 18:30:18.918551 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4\": container with ID starting with 3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4 not found: ID does not exist" containerID="3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.918631 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4"} err="failed to get container status \"3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4\": rpc error: code = NotFound desc = could not find container \"3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4\": container with ID starting with 3991b4c0d3055129ba362b8d27ec39ce092fe3283477642f66c0124efd0acdb4 not found: ID does not exist" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.918676 4809 scope.go:117] "RemoveContainer" containerID="a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6" Nov 27 18:30:18 crc kubenswrapper[4809]: E1127 18:30:18.919086 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6\": container with ID starting with a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6 not found: ID does not exist" containerID="a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.919115 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6"} err="failed to get container status \"a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6\": rpc error: code = NotFound desc = could not find container \"a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6\": container with ID starting with a5482405a5cfc13b24bf73350e27ad22acca78699da86df4b89ed0e4cfeb3ca6 not found: ID does not exist" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.919136 4809 scope.go:117] "RemoveContainer" containerID="7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482" Nov 27 18:30:18 crc kubenswrapper[4809]: E1127 18:30:18.919670 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482\": container with ID starting with 7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482 not found: ID does not exist" containerID="7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482" Nov 27 18:30:18 crc kubenswrapper[4809]: I1127 18:30:18.919704 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482"} err="failed to get container status \"7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482\": rpc error: code = NotFound desc = could not find container \"7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482\": container with ID starting with 7a8515b3390ba1c5ee098b2ae17f16e2e054c113c0a934c9b2addb590b8d3482 not found: ID does not exist" Nov 27 18:30:19 crc kubenswrapper[4809]: I1127 18:30:19.492616 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="006fbfe2-b2e2-4707-aa60-898e8feb6e76" path="/var/lib/kubelet/pods/006fbfe2-b2e2-4707-aa60-898e8feb6e76/volumes" Nov 27 18:30:21 crc kubenswrapper[4809]: I1127 18:30:21.459423 4809 scope.go:117] "RemoveContainer" containerID="5ac7ef6a3ecc87f6e860f3d3c64e22b39db2a069221513c82a2608d075142779" Nov 27 18:30:21 crc kubenswrapper[4809]: E1127 18:30:21.461149 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qwx9w_openshift-machine-config-operator(6b63ded2-aa4f-4aba-b3cc-a3965a01036c)\"" pod="openshift-machine-config-operator/machine-config-daemon-qwx9w" podUID="6b63ded2-aa4f-4aba-b3cc-a3965a01036c" Nov 27 18:30:21 crc kubenswrapper[4809]: E1127 18:30:21.826260 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice/crio-f4c188370caf5c64f579c376e81553e88967f5bc241aaf3205d1fb7c09b1a72b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea3dc648_304d_430a_87a1_5a3c315ad238.slice\": RecentStats: unable to find data in memory cache]" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112114315024437 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112114315017354 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112102366016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112102366015452 5ustar corecore